Abstract
Introduction
Alzheimer’s disease is a progressive neurodegenerative disorder challenging early diagnosis and treatment. Recent advancements in deep learning algorithms applied to multimodal brain imaging offer promising solutions for improving diagnostic accuracy and predicting disease progression.
Method
This narrative review synthesizes current literature on deep learning applications in Alzheimer’s disease diagnosis using multimodal neuroimaging. The review process involved a comprehensive search of relevant databases (PubMed, Embase, Google Scholar and ClinicalTrials.gov), selection of pertinent studies, and critical analysis of findings. We employed a best-evidence approach, prioritizing high-quality studies and identifying consistent patterns across the literature.
Results
Deep learning architectures, including convolutional neural networks, recurrent neural networks, and transformer-based models, have shown remarkable potential in analyzing multimodal neuroimaging data. These models can effectively process structural and functional imaging modalities, extracting relevant features and patterns associated with Alzheimer’s pathology. Integration of multiple imaging modalities has demonstrated improved diagnostic accuracy compared to single-modality approaches. Deep learning models have also shown promise in predictive modeling, identifying potential biomarkers and forecasting disease progression.
Discussion
While deep learning approaches show great potential, several challenges remain. Data heterogeneity, small sample sizes, and limited generalizability across diverse populations are significant hurdles. The clinical translation of these models requires careful consideration of interpretability, transparency, and ethical implications. The future of AI in neurodiagnostics for Alzheimer’s disease looks promising, with potential applications in personalized treatment strategies.
Keywords: Alzheimer’s disease, deep learning, multimodal neuroimaging, Disease progression prediction, brain imaging analysis
Graphical abstract
Highlights
Integration of MRI, PET, and fMRI with deep learning models for early AD detection.
Exploration of CNNs, RNNs, and transformers in analyzing brain imaging data.
Use of AI to predict disease progression, differentiate MCI and AD stages, and identify biomarkers.
Addressing issues like data heterogeneity, small sample sizes, and generalizability.
Discussing model interpretability, regulatory considerations, and AI’s role in personalized AD treatments.
1. Introduction
1.1. Overview of Alzheimer’s disease and the importance of early diagnosis
Alzheimer’s Disease (AD) is defined as a neurodegenerative disorder characterized by cognitive decline, memory loss, and behavioural changes which cannot be reversed (Porsteinsson et al., 2021). There is no cure for this disease but treatments are available which can decline its progression (Raza et al., 2024a; Atri, 2019). It is one of the most common causes of dementia, affecting 50% of the 85-year-old age group population (Aramadaka et al., 2023). AD causes the degeneration of neurons (Raza et al., 2024b), specifically those that are involved in memory and intellectual functions (Atri, 2019).
It’s very important to diagnose AD before its onset. While there is no cure for AD, early diagnosis allows health professionals to take timely intervention (Raza et al., 2024a) and screening to manage the disease so that its progressive nature can be slowed down (Porsteinsson et al., 2021). It can also help patients as well as care providers in a way that they can plan for the necessary lifestyle changes in future to make the patient’s life quality better (Porsteinsson et al., 2021). According to researchers, rolling back the diagnosis by 5 years could reduce the frequency of AD by 50% (Amira et al., 2015). Challenges have often come along with the early diagnosis, which includes limitation of time for clinicians, proper diagnosis of AD pathology and misdiagnosis of the symptoms, which are similar to the normal ageing process (Porsteinsson et al., 2021).
Despite facing such challenges, early diagnosis and continuous research led to the development of advanced diagnostic tools and biological markers. Many diagnostic tests have been investigated, such as neuropsychological and neuroimaging tests, and their working with machine learning (ML) and deep learning (DL) showed promising results in diagnosing AD at its early stages (Ding et al., 2019; Feng et al., 2020). The use of Electroencephalography (EEG) together with ML has shown high accuracy in diagnosing AD patients (Pirrone et al., 2022). Neuropsychological testing focuses on dual-task performance and active inhibition to catch AD in its early stages while differentiating it from the control group (Toepper et al., 2008).
Magnetic Resonance Imaging (MRI) and Computer-Aided Diagnostic (CAD) systems, which are classified under Neuroimaging techniques, were found to achieve good accuracy in the early diagnosis of AD. MRI and CAD use Deep Neural Networks (DNN) and Convolutional Neural Networks (CNNs), which are advanced subfields of ML; by using these approaches, MRI and CAD detect AD-related patterns with accuracy higher than 90% (Shah et al., 2024; Amira et al., 2015; Rabeh et al., 2023).
1.2. Role of neuroimaging in Alzheimer’s disease diagnosis
Neuroimaging is defined as the process of image production of the structure or the activity of the central nervous system, primarily the brain, through techniques like MRI or computerized tomography (CT). However, neuroimaging plays an important role in diagnosing AD. Positron emission tomography (PET) and single-photon emission computed tomography (SPECT) are the techniques which have been useful in detecting structural and functional changes in the brain that are associated with AD. These techniques have greatly contributed to AD research and diagnosis (Weinstein et al., 1993).
MRI can detect fine structural and functional changes in different regions of the brain, including the integrity of the blood–brain barrier, through dynamic contrast-enhanced MRI (DCE-MRI) technique (Raja et al., 2018). MRI has been widely used in AD diagnosis. Brain atrophy patterns, which are the characteristics of the AD process, can be detected through volumetric MRI (Zhou et al., 2024). Neuroimaging data has a multi-dimensional nature; this can present analytical issues (Shu, 2016). To address this, researchers have developed voxel-level analyses, region of interest (ROI) studies and more advanced procedures like seed-voxel and multivariate techniques (Albajes-Eizagirre et al., 2019; Habeck et al., 2010). These techniques can assist in extracting relevant features and patterns from brain images.
The emergence of AI and advances in ML have transformed the analysis of neuroimaging in diagnosing AD. These advanced computational tools give more precise diagnoses of AD patients and disease progression. Deep learning method CNNs extract relevant features from brain images while classifying them into different stages of cognitive impairment (Parmar et al., 2020). However, this association of AI and ML provides diagnostic accuracy and efficient analyses of radiographic data and contributes to the development of precision medicine approaches for AD (Mirkin and Albensi, 2023; Jo et al., 2019).
Longitudinal data analysis of neuroimaging data is another important aspect of AD discussion. It allows for the study of disease progression over time (Aberathne et al., 2023). It can also help in identifying vulnerable brain regions and determining threshold values for biomarkers such as plaques, tangles, and neurodegeneration (Aberathne et al., 2023). The diverse effects of AD on the brain have been studied by employing normative modelling techniques. These techniques analyze specific individual changes across multiple imaging biomarkers that reveal the effect of AD on the brain (Verdi et al., 2024). While neuroimaging biomarkers have shown promise in AD diagnosis, they are often limited in clinical application due to their high cost and, in some cases, invasive nature (Khan and Alkon, 2015). Research is needed for more accessible and cost-effective biomarkers, such as blood-based markers, which can introduce neuroimaging techniques in low and middle-income healthcare settings.
Neuroimaging when associated with AI and ML can be a great tool for the early and more accurate diagnosis of AD as It provides valuable insights into brain structure and function that helps in monitoring disease progression. Research is needed to deal with the challenges in standardizing methods and validating findings across different populations.
1.3. Advancements in deep learning for brain imaging analysis
Deep learning (DL) is a type of machine learning (ML) that utilize artificial neural networks having multiple layers of processing and extracting high level features from the data while performing complex task like classification, regression and representation of the data. DL has made great advancements in brain imaging analysis serving the field of neuroscience and medical imaging.
Artificial neural networks such as CNNs, which are the type of DL algorithm used to identify patterns in images and perform computer vision tasks, showed success in diagnosing and classifying the imaging data for the early detection and classification of AD (Wen et al., 2020). CNNs can be applied to MRI and PET images for the identification and classification of different structures enhancing the accuracy of diagnosis of AD (Wen et al., 2020). It provides great help for the radiologist to interpret complex cases more effectively and decrease their workload through its automation process. However, the circle of DL is not limited to just AD. It extends far beyond it. It works with several other brain disorders, such as Parkinson’s disease, autism spectrum disorder, and Schizophrenia (Zhang et al., 2020a).
DL methods have bypassed the traditional ML techniques in diagnosis through computer-aided techniques, and generative adversarial networks (GANs) have been explored to overcome the limitations associated with data availability in brain imaging studies (Galić et al., 2023; Logan et al., 2021). DL methods, when compared to standard ML methods, were found to have a better-scaling potential and their ability to make the most of nonlinearities in neuroimaging data to make more effective representations of tasks for better characterization of the human brain (Abrol et al., 2021). From adult to pediatric medical imaging, DL has been instrumental in detecting neurological developmental disorders by identifying brain and tissue structures while predicting the outcomes (Hu et al., 2023).
However, DL approaches have greatly covered several aspects that include brain imaging, early detection of disorder, classification and data representation but further research is needed to address the interpretation issues and medical images variability to utilize the full potential of DL.
1.4. Objectives of the review
To provide an overview of AD and emphasize the significance of early diagnosis in improving patient outcomes.
To explore the role of neuroimaging techniques, such as MRI and PET, in the diagnosis and monitoring of AD.
To review advancements in deep learning methodologies applied to brain imaging analysis, particularly for the early detection and progression of AD.
To discuss various deep learning architectures, including CNNs, RNNs, transformers, and hybrid models, and their applications in structural and functional neuroimaging for AD diagnosis.
To evaluate the role of MRI and PET fusion and other structural imaging and functional imaging techniques in providing early diagnosis of AD.
To analyze the use of deep learning methods in predictive modelling to track disease progression, find biomarkers and distinguish MCI from AD.
To underline the issues in neuroimaging data standardization, including variability in imaging protocols, limited data availability, and generalizability across diverse populations.
To provide suggestions for future research, such as the development of explainable deep learning models, multimodal data integration, and solutions for overcoming data-related limitations.
2. Convolutional neural network for structural imaging analysis
CNNs are the cornerstone of deep learning models capable of handling high-dimensional data such as neuroimages. Its basic framework consists of an input layer, some middle layers, and an output layer. The hidden layers include the convolutional layers (C-layers), the subsampling layers (S-layers), and the fully connected layers (FC-layers). All these layers cooperate to identify the valuable insight in the data, to decrease the data volume, and to categorize the data.
As the CNNs advance, they become refined and incorporate new changes in the imaging equipment, such as MRI and PET scans. However, when trained on larger data sets, it is boosted to a great extent. The fact that CNNs can learn directly from the images means that the features can be learned automatically without having to handcraft them. This potential makes them suitable for various applications, making them effective in measuring the accurate staging of Alzheimer’s disease.
In AD studies, CNNs are trained to recognize morphological and metabolic alterations of the brain by concentrating on specific areas, including the hippocampus, amygdala, and insula. Their capability to learn from these areas of the damaged brain makes them crucial for early-stage detection and diagnosis. For instance, CNNs have the potential to recognize hippocampal and cortical atrophy from MRI or abnormalities in glucose metabolism and amyloid deposition from PET. Due to the ability of CNNs to handle massive high-dimensional data sets such as the ones from ADNI and OASIS, which contain neuroimaging and clinical information, CNNs are employed.
CNNs can enhance the accuracy of detecting Alzheimer’s even before the appearance of clinical symptoms by learning from a lot of data. They also help in the creation of predictive models as well as individualized interventional approaches. These deep-learning models are more convenient for medical images since they can learn hierarchical features (Saleem et al., 2022). Some of the CNN models that have been employed in the diagnosis of AD include AlexNet and DenseNet, among others. These models not only distinguish between a healthy person and a patient but can also assess the patient’s condition as in either the early stage of the disease or the advanced stage, incorporating the mild cognitive impairment (MCI) and the early stage of AD. For instance, AlexNet has exhibited remarkable results in analyzing MRI images and thus set a high standard in this field (Dakdareh and Abbasian, 2024).
2.1. Recurrent neural networks (RNNs) for longitudinal data and temporal analysis
Recurrent Neural Networks (RNNs) could be considered one of the best approaches to processing longitudinal and temporal physiological data because of their capability to manage sequential inputs and temporal dependencies. DNNs, CNNs, or other machine learning models could be used in analyzing dynamic systems, but RNNs are best suited for this purpose. The features they consider are both the current and previous sequences of features, which is useful when faced with input sequences of varying lengths (Mao and Sejdic, 2023).
In the case of MCI to AD conversion, annual NP test data are used by RNNs to incorporate temporal patterns of the disease. This allows RNN models to capture the change from MCI to AD over several years and indicates the patient’s future prognosis. It can be concluded that the proposed RNN-based models are more effective than traditional models (Table 1), including logistic regression, in the case of using fewer features and without using expensive data such as MRI images. This can make a difference in the quality of the patient and the cost of the treatment; in fact, it has the potential to place RNN-based predictions as the new generation of Alzheimer’s treatment.
Table 1.
Comparative analysis of deep learning and traditional models.
Approach | Examples | Accuracy (%) | AUC | Key Features | Reference |
---|---|---|---|---|---|
Traditional ML | SVM, Random Forest, Decision Trees | 94.5% | 0.96 | Requires manual feature extraction; interpretable; data-efficient | Alshamlan et al. (2024) |
Deep Learning (CNNs, RNNs, Transformers) | AlexNet, DenseNet, Vision Transformer, 3D-CNN | 94–99 | 0.91–0.99 | End-to-end learning; automatic feature extraction; scalable for large datasets | Alsubaie et al. (2024); Khan and Saddam (2024) |
Hybrid DL Models | CNN + RNN, Transformer-CNN Fusion | 91–99 | 0.937 | Multimodal learning; enhanced accuracy by combining imaging and clinical data | Asaduzzaman et al. (2025); Vashishtha et al. (2023) |
Contrary to other methods that depend on costly and detailed imaging studies, RNNs use easily accessible NP tests to make relevant and timely predictions based on patients’ records. In this manner, RNNs can predict the stages of MCI to AD over the next few years, which will also be more inexpensive and less invasive than the current diagnosis and management of the disease (Park et al., 2024) In their present form, RNNs could be more effective than traditional models like logistic regression, even with few input data. This advancement can also improve the lives of the patients and the healthcare system, which results in a decline in costs for the treatment of AD (Al Olaimat and Bozdag, 2024).
2.2. Transformer and attention mechanisms for multimodal data integration
Transformers have played a crucial role in the development of AI, especially in dealing with sequential data using innovative approaches to the self-attention mechanism. This mechanism allows the transformers to work on the parts of a sequence that are relevant to the task at hand, irrespective of their location Table 2. Therefore, the models are capable of learning long-range dependencies (Madan et al., 2024). The transformers were initially designed for natural language processing and have been adopted in healthcare, especially in such areas as the diagnosis. AD is complex and therefore has multiple etiologies, and the progression of this disease needs several data sets for better prediction.
Table 2.
Quantitative performance metrics for deep learning models in Alzheimer’s diagnosis.
Model | Modality | Accuracy (%) | Sensitivity (%) | Specificity (%) | AUC | Reference |
---|---|---|---|---|---|---|
CNN (AlexNet, DenseNet, ResNet) | MRI | 89.4–98.9 | 86.7–98.9 | 84–98.8 | — | Feng et al. (2020) |
3D ResAttNet | MRI | 84–93.44 | — | — | 81.8–96.7 | Wang Y. et al. (2024) |
RNN (LSTM, GRU-based) | Longitudinal Data | 75.57 | — | — | — | Airlangga (2024) |
Transformer (Vision Transformer, Med-PLM, BRLTM) | Multimodal Data | 92.14% | 93.27% | 89.95% | — | Zhao et al. (2024) |
Hybrid Models (CNN + RNN, HiLCAE +3D-VGG16, BGRU + DenseNet) | MRI + PET | 98.8% | 100% | 76% | — | Khatun et al. (2023) |
In the framework of Alzheimer’s diagnosis, transformers exhibit tremendous potential in aggregating various types of data such as EHRs, genetics, and imaging. Transformers are especially effective in dealing with multimodal data since Alzheimer’s presents differently in every patient. This includes data, in the form of brain scans and genetic markers, which offer a more complete and specific view of the patient. For instance, some models like Med-PLM and BRLTM have incorporated diagnosis codes, medical histories, and even unstructured clinical notes to enhance the diagnostic abilities of the model, which shows the future of transformers in this domain.
Additionally, transformers are getting more recognition in medical imaging, especially Vision Transformers (ViTs), as they have the potential to handle larger effective receptive fields. This makes them capable of capturing more contextual information in medical images than the conventional CNNs; this is very important as it determines the overall diagnosis of the tissue and organs. Although they face difficulties in terms of computational intensity and the requirement of big data sets, transformers are being considered a great perspective in medical imaging as a potential way to improve the quality of healthcare, especially in the context of Alzheimer’s diagnosis (Tang et al., 2023).
2.3. Hybrid deep learning model combining multiple techniques
The hybrid deep learning models have enhanced the performance of diagnosing AD to the state where it has not been before through a combination of different architectures that increase the accuracy and reliability of the model. These models incorporate various models such as CNNs, Auto-Encoders (AEs), and RNNs to process multi-modal data, including MRI and PET images as well as speech samples (Saleem et al., 2022).
For instance, the 3D Explainable Residual Self-Attention Convolutional Neural Network (3D ResAttNet) uses a self-attention mechanism along with residual connections to improve attention towards important features while at the same time improving the training efficiency. Such models as sparse autoencoders (AEs) and capsule networks are applied for the refinement of feature extraction with the purpose of enhancement of detection. Further, the high-level layer concatenation autoencoder (HiLCAE) along with 3D-VGG16 has been used to improve diagnostic performance by using both MR and PET images. To examine the feasibility of AD, CNNs and RNNs are used to analyze speech information (Saleem et al., 2022).
Further, Bidirectional Gated Recurrent Units (BGRUs) combined with Dense Convolutional Networks (DenseNets) enable us to learn spatial and temporal patterns from the brain data. Other complicated structures like SBi-RNNs combined with 3D-CNNs, as well as Dense Connection CNNs with attention, have, however, been indicated to perform better than the traditional methods. These models can therefore be used to improve and expedite the diagnosis of AD, making the models relevant in any clinical setting. Nationally, models such as Bidirectional Gated Recurrent Units (BGRU) along with Dense Convolutional Networks (DenseNets) are capable of learning both spatial and temporal features from brain data. Other complex architectures, including Stacked Bidirectional Recurrent Neural Networks (SBi-RNNs) incorporated with 3D-CNNs and Densely Connected Convolutional Neural Networks (DCCNNs) with attention mechanisms, have been shown to outperform the conventional approaches (Saleem et al., 2022). These models provide a platform for enhanced and timely AD diagnosis, thus making them applicable in any given clinical setting.
One such model is the 3D Hybrid Compact Convolutional Transformer (3D HCCT), which is an advanced model for the AD classification from 3D MRI scans. These networks are convenient in learning the local features, while these are optimal when our focus is on the global forms and the long-term relations. The entity acquires the local features while ViTs are effective at capturing the global patterns and long-range dependencies. This architecture combines the two models to produce richer feature representations that are more discriminative and therefore achieve better accuracy, and sensitivity, in contrast to the traditional CNN-based models, specifically when tested on the Alzheimer’s Disease Neuroimaging Initiative (ADNI) data sets (Majee et al., 2024).
Also, the 3D HCCT’s ability to perform well on various data sets and its efficient architecture that reduces the consumption of resources makes it suitable for use in the clinical setting. Its ability to aid in the diagnosis of AD at an early stage and with a high level of accuracy shows how AI in healthcare can transform the healthcare industry and improve patient care.
3. Integration of structural and functional imaging
The fusion of resting-state functional MRI (rs-MRI) with Structural MRI has shown better outcomes in the early detection as well as staging of AD (Hojjati et al., 2018) the multimodal approaches demonstrates great promise in accuracy as compared to single modality approaches. For instance, combined sMRI and rs-fMRI modality outperformed sMRI (89%) and rs-fMRI (93%) alone and demonstrated 97% accuracy in diagnosing MCI converters against non-converters (Hojjati et al., 2018).
Likewise, combining MRI, fMRI, and cognitive test scores produced 95.65% accuracy in distinguishing AD patients from healthy controls (Dachena et al., 2020). Though, the advantages of combining functional and structural imaging can vary across patient categories. For example, integration improves differentiation between AD and behavioral variant frontotemporal dementia (bvFTD) but does not always significantly outperform structural measures alone when distinguishing AD from healthy controls (Bouts et al., 2018) Therefore, multimodal imaging approaches effectively capture complementary information about brain structure and function, enhancing diagnostic accuracy and strength. Additional research is required to optimize feature selection, integration methods, as well as clarification of the role of each modality across different patient populations.
3.1. MRI and PET fusion for early detection of Alzheimer’s disease
The combination of MRI and PET methods demonstrates better outcomes in the early detection of AD by synchronizing the structural revelations of MRI to the metabolic status captured by PET. This integration results in a more comprehensive view of brain changes. Studies have revealed that multimodal fusion approaches outperform individual modality methods in diagnostic accuracy (Castellano et al., 2024; Du et al., 2024a; Shukla et al., 2023). Noteworthy methods embrace the integration of gray matter (GM) areas from MRI with FDG-PET (Fluorodeoxyglucose-positron emission tomography) to create a “GM-PET” modality and a wavelet transform-based fusion process combined with deep learning for categorization. These techniques have attained up to 99% accuracy in binary classifications (Alzheimer’s disease AD vs. Cognitively Normal and Mild Cognitive Impairment vs. Cognitive Normal) and 96% in multi-class classifications (AD vs. MCI vs. CN), highlighting their potential for strong and early AD diagnosis (Parihar and Swami, 2024; Song et al., 2021) By combining structural and functional imaging data, these advancements support former interventions and better patient outcomes (Shukla et al., 2023) Table 3 depicts the comparison of single modalities with fusion modalities.
Table 3.
Comparison of single modality (MRI, PET) and fusion approach (MRI-PET Fusion) for the early diagnosis of Alzheimer’s Disease.
Modality | Study Findings | Evidence |
---|---|---|
MRI | MRI is effective in detecting brain atrophy and volumetric changes, specifically in the hippocampus and medial temporal lobe. | Rao et al. (2023) |
Demonstrates high sensitivity in detecting early structural changes in gray matter but limited specificity in differential diagnosis. | Chouliaras and O’Brien (2023) | |
PET | FDG-PET detects early metabolic changes in the posterior cingulate and precuneus regions, preceding structural atrophy. | Mosconi et al. (2004) |
PET using amyloid tracers (e.g., PiB) improves the detection of amyloid-beta plaques in preclinical stages of Alzheimer’s. | Suppiah et al. (2019) | |
MRI-PET Fusion | Combining PET and MRI enhances the detection of early biomarkers by integrating structural and metabolic changes. | Lee et al. (2024). |
Integration of PET and MRI data improves diagnostic accuracy through automated machine learning techniques. | Castellano et al. (2024) | |
Demonstrates improved sensitivity and specificity in early detection by combining anatomical and functional data. | (Song et al., 2021) | |
Fusion models using deep learning enhance early detection with improved performance compared to single modalities. | Venugopalan et al. (2021) | |
Advanced neural networks combining MRI and PET achieve higher diagnostic accuracy in distinguishing Alzheimer’s from other dementias. | Huang et al. (2019) |
The following Table 4 enlightens different multimodal fusion approaches combining MRI and PET imaging for the early diagnosis of AD.
Table 4.
Multimodal fusion of MRI and PET imaging for detection of Alzheimer’s Disease (AD).
Fusion Technique | Description | Key Findings | Reference |
---|---|---|---|
Image-level Fusion | Combines MRI and PET images to create a single composite image. Example: “GM-PET” modality fuses gray matter tissue area of MRI with FDG-PET images. | Retains both structural and metabolic information; outperformed unimodal and feature fusion methods. | Song et al. (2021) |
Feature-level Fusion | Extracts features from MRI and PET separately, then combines them. Example 1: Wavelet transform-based fusion with deep learning using ResNet-50. | Incorporates structural and metabolic data effectively for improved diagnosis. | Choudhury et al. (2024) |
Example 2: Pre-trained neural networks for 2D feature fusion of MRI and PET. | Enhanced feature extraction and fusion capabilities. | Suma et al. (2022) | |
Decision-level Fusion | Classifiers are trained on individual modalities (MRI, PET, EEG), and outputs are combined. | Ensemble classifier approach improved diagnostic accuracy by 10–20% compared to individual modalities. | Polikar et al. (2010) |
3D Fusion Techniques | Operations include skull-stripping, image segmentation, and co-registration. Example: Comparison of 2D and 3D CNN architectures for MRI-PET fusion. | 3D fusion achieved good accuracy (86.90%) for AD classification; 3D methods showed effectiveness. | Suma et al. (2022) |
3.2. Functional connectivity and resting-state fMRI in AD diagnosis
Resting-state functional magnetic resonance imaging (rs-fMRI) has emerged as a promising tool for studying Alzheimer’s disease (AD) and its preclinical stages. Various studies have evidences that rs-fMRI can differentiate functional connectivity changes in the brain’s default mode network (DMN) across the spectrum of AD, from healthy aging to MCI and clinical AD (Binnewijzend et al., 2012; Damoiseaux, 2012). These variations are often characterized by diminished connectivity within the DMN, predominantly in the precuneus and posterior cingulate cortex, and are not dependent on cortical atrophy (Agosta et al., 2012; Binnewijzend et al., 2012) In the course of decreasing DMN connectivity in AD, some studies have detected enhanced connectivity in frontal networks, probably as a compensatory mechanism to preserve cognitive efficiency (Agosta et al., 2012).
Furthermore, AD patients have shown altered connectivity patterns between different brain areas, such as diminished positive correlations between prefrontal and parietal lobes and increased positive correlations within lobes (Ibrahim et al., 2021). These findings support the concept of an anterior–posterior disconnection phenomenon in AD. Thus rs-fMRI holds potential as a biomarker for monitoring of early detection and progression of disease. However, standardization of methodologies and analysis techniques is required for clinical implementation. Integrating rs-fMRI with other imaging modalities and cognitive measures may further improve diagnostic accuracy and provide deeper insights into AD pathophysiology (Damoiseaux, 2012).
3.3. Multimodal data preprocessing and alignment techniques
Multimodal data preprocessing and alignment techniques play a key role in the diagnosis of AD using neuroimaging data. Various studies have emphasized the importance of these modalities in refining diagnostic accuracy and elucidating meaningful features from different imaging modalities. Techniques like age correction, feature selection, and feature reduction enable meaningful analysis, with these methods enhance diagnostic accuracy by normalizing and reducing the dimensionality of data from various modalities, such as MRI, PET (positron emission tomography), CSF (cerebrospinal fluid) biomarkers, and genetic features (Lin et al., 2021). Some studies also focus on specific brain regions like the hippocampus for extracting discriminative features (Ahmed et al., 2019).
Though integration of multimodal data offers significant advantages, integrating high-dimensional imaging and genetic data poses challenges (Peng et al., 2019). To address these challenges, new approaches like structured sparsity and multiple kernel learning methods have been introduced to facilitate structured feature selection and fusion from mixed modalities. In conclusion, advanced preprocessing and alignment techniques, such as graph diffusion methods for enhancing similarity measures among subjects (Wang et al., 2023) and image fusion methods combining gray matter tissue areas of brain MRI and FDG-PET images (Song et al., 2021) have demonstrated improved performance in AD diagnosis. These advanced techniques improve data quality and integration, resulting in more precise and explainable diagnostic models for AD.
3.4. Challenges in integrating heterogeneous imaging data
The diverse range of imaging techniques, including CT, MRI, and PET, each offer unique understandings into AD pathophysiology but also poses complications in data integration (Johnson et al., 2012). These modalities provide varying types of information, from structural alterations to functional changes and molecular pathology, making it problematic to create a unified understanding of the disease process and progression. One significant challenge is the apparent paradox between the clinically diverse presentations of AD and the common underlying histopathological substrate.
One significant challenge is the apparent paradox between the clinically diverse presentations of AD and the common underlying histopathological substrate. Variant syndromes of AD, such as posterior cortical atrophy and logopedic variant primary progressive aphasia, exhibit different clinical phenotypes despite sharing similar pathological pictures (Warren et al., 2012). This heterogeneity complicates the interpretation of imaging data and highlights the need for a more refined understanding of how different brain networks are affected in AD (Loreto et al., 2024). To address these challenges, researchers are developing comprehensive knowledge bases like the Alzheimer’s Knowledge Base (AlzKB), which combines diverse data sources to provide a complete depiction of AD etiology and potential therapeutics (Romano et al., 2024).
Such approaches aim to capture the complex relationships between biological and pharmaceutical entities at various levels of organization. However, the successful integration of heterogeneous imaging data will require continued efforts to standardize data collection, develop advanced analytical techniques, and create frameworks that can accommodate the multifaceted nature of AD pathology and its clinical manifestations. Table 5 highlights the challenges in integration of heterogenous data.
Table 5.
Challenges in integrating heterogeneous imaging data.
Challenges | Description |
---|---|
Diverse Insights from Multiple Imaging Modalities | CT (Computed Tomography): Detects structural abnormalities and excludes other causes of cognitive decline, such as tumors or significant brain atrophy. MRI (Magnetic Resonance Imaging): High-resolution imaging to detect hippocampal atrophy, a hallmark of AD. PET (Positron Emission Tomography): depicts molecular /functional data, e.g., glucose metabolism or amyloid and tau deposits. Each modality, i.e., CT, MRI and PET provides unique understandings but poses challenges for integration because of different spatial/temporal resolutions and biological targets (Zhang Y.-D. et al., 2020) |
Challenges in Data Integration | Heterogeneity of data: Structural changes (e.g., atrophy) occur late, while functional/molecular changes manifest earlier but are harder to contextualize (Hofmann-Apitius et al., 2015) Variability in protocols and biomarkers: Differences in acquisition methods complicate the merging of diverse datasets (Nazir et al., 2024) |
Clinical and Pathological Paradox | AD’s clinical diversity (e.g., posterior cortical atrophy, logopenic variant primary progressive aphasia) contrasts with its uniform pathology (amyloid plaques, tau tangles), creating challenges in linking imaging findings to varied clinical presentations (Jellinger, 2022) |
In order to minimize the complexity of heterogenous data integration, additional researches are required to be directed for standardization of data collection methods across institutions and studies (Mueller et al., 2005). This will facilitate comparability and reproducibility of data. Furthermore, development of sophisticated analytical methods capable of combining data from multiple modalities (e.g., MRI-PET fusion) needs to be developed (Castellano et al., 2024). It is also required to create frameworks for data interpretation that can accommodate AD’s complexity and variability, incorporating clinical, imaging, genetic, and biomarker information (Li et al., 2015). Therefore, the integration of diverse imaging data in AD research holds the promise of providing a more holistic understanding of the disease. However, achieving this goal requires continued innovation in data standardization, analytical techniques, and interdisciplinary collaboration to address the multifaceted nature of AD pathology and its clinical presentations.
4. Predictive modeling of disease progression
Disease progression modelling is currently a key asset in most medical practices given that it assists clinicians and researchers to predict the course of a disease and respond adequately on preventive measures, diagnosis and treatment (Alowais et al., 2023). In particular, these models employ statistical and machine learning approaches to gather an enormous amount of data, describe its distribution, and forecast potential further events for specific patient or multiple patients at once (Raza and Abbas, 2024; Babb de Villiers et al., 2020). Of all the chronic and progressive diseases, they are highly essential for several diseases like cancer, cardiovascular diseases, diabetes and neurodegenerative diseases where timely intervention can greatly determine the state of the patients (Raza and Abbas, 2024; Wang J. et al., 2024).
It is a machine learning model, for instance, random forests, support vector machines (SVMs), and deep neural networks which deal with the data sourced from Electronic Medical Records, Imaging scans, Genomic data, wearable devices with a high degree of certainty (Raza et al., 2024b; Raza and Abbas, 2024; Du et al., 2020). For instance, in neurodegenerative diseases such as Alzheimer’s predictive models evaluates clinical and imaging biomarkers, longitudinal data on cognition and brain imaging to determine risk of further deterioration or progression from probable MCI to dementia (Du et al., 2024b). Temporal modelling is a type of predictive modelling that focuses on time series, disease trajectory (Bilgel et al., 2017).
Functions like RNNs and long short-term memory networks (LSTMs) are highly adept at identifying variations, and therefore are suitable for conditions that have over time varying biomarkers or symptoms (Othman et al., 2024). For instance, temporal models are used to forecast the growth rate of the tumor in cancer or blood sugar levels in diabetes to allow the clinician to prevent emergent complications on the basis of treatment (Sharma et al., 2017). Risk assessment is another of the most important applications of predictive modeling, which sorts patients by their risk level or the degree of risk that they experience adverse outcomes (Vadapalli et al., 2022). There is always some stratification to help identify patients for closer observation or immediate action. In addition, predictive models can accommodate patient-specific characteristics, including genetic attributes or lifestyle, and determine specific interventions possible for the patient with the disease (Tong et al., 2024).
Nonetheless, predictive modeling has some limitations such as data quality problems, a shift in dataset distribution, or high dimensionality, and requirements for sample size and variability (Lombardo et al., 2024). However, the interpretability of the learned model needs to be guaranteed for it to be useful in the clinical settings. Due to the growing complexity of these models, to make them more understandable and to build trust in such models, Explainable AI, or XAI, methodologies are being researched (Wyatt et al., 2024). Subsequently, as the field of predictive modelling progresses, the capability of incorporating online data from wearable devices and digital health applications will expand the usability of the model to become a more proactive field of healthcare.
4.1. Deep learning for predicting disease trajectory and biomarker identification
Therefore, among the most revolutionary instruments of early disease trajectory prediction and prognostic biomarker discovery is made possible by deep learning (DL), which falls within the field of artificial intelligence (Miller et al., 2024). AI will play a vital role in deciphering the mechanisms of disease, forecasting disease progression, and developing targeted therapies, thanks to its ability to analyse extensive and complex datasets rapidly and autonomously (Abbas et al., 2019; Zhang and Shen, 2012).
Deep learning extracts informative patterns from multimodal data such as medical imaging, genomics, and electronic health records (EHR) by leveraging more advanced architectures such as CNN, RNN, and Transformers (Li et al., 2020). Deep learning models use longitudinal data to analyze patterns and make predictions about future health trajectories (Du et al., 2020). As for neurodegenerative diseases such as Alzheimer’s, linear convolutional networks (LCNs) can manage brain imaging data to uncover early structural changes, while RNNs, for analysis of temporal data (for instance, cognitive scores and levels of biomarkers) can be used to predict disease progression by capturing the temporal dynamics (Alsubaie et al., 2024). They are also being applied to diabetes treatment, where deep learning models analyze continuous glucose monitoring data to predict changes in blood sugar levels, allowing for early medical interventions (Faruqui et al., 2019). Biomarkers have become increasingly essential to diagnostics and targeted treatments, and for the successful identification, deep learning has also emerged as a key parameter (Echle et al., 2021).
Deep learning models can derive potential biomarkers from genomic, proteomic, or metabolomic data that indicate the onset or progression of a disease event (Mann et al., 2021). For example, autoencoders and deep generative models may capture genomic variations present but often overlooked by DNA sequencing (e.g., somatic mutations in cancer) or less frequently examined patient specimens (e.g., rare syndromes), CNNs apply deep learning to histopathological images of solid tumors to examine cancer-specific markers (Guleria et al., 2023).
The fusion of multimodal data also demonstrates the deep learning ability across single/multiple factors (Li et al., 2022). Models trained on imaging plus genetics plus whatever clinical records can discover complex relationships and interactions, delivering a holistic view of disease dynamics (Kuznetsov et al., 2013). This is particularly useful for complex diseases such as cancer and autoimmune diseases, where altered pathways contribute to disease progression (Guleria et al., 2023). Medical deep learning brings the advantages but also challenges. This requirement of large labeled datasets can often be a bottleneck in training a model, and bias in the data can impact generalizability (Paproki et al., 2024).
Another issue is interpretability since most of deep learning models are “black boxes” and it is not easy to understand how the predictions are made (Paproki et al., 2024). To overcome these challenges there is required a joint effort of data scientist, clinicians, and regulatory bodies to ensure that models are indeed accurate, reliable, and ethical (Guan, 2019). The progress has been sustained in understanding the disease prognosis and biomarker discovery by deep learning enabled through computing capabilities and availability of data. When applied to health care, its uses promise early diagnoses, right treatment, and better results offering directions to precision medicine.
4.2. Temporal modeling for tracking cognitive decline
Temporal modeling is a particularly effective approach for tracking the specific cognitive changes that define AD, Parkinson’s disease, and other forms of dementia (Kunst et al., 2019). Temporal models come in handy in analyzing longitudinal data in that, they can identify change in cognitive function, biomarkers, or many other clinical assessments with time. They are also useful when it comes to diagnosing conditions at their early stage, as well as tracking the progression of a disease, and delivering personalized treatment (Zuidersma et al., 2019).
The modeling in this context involves adoption of various machine learning techniques especially the RNNs and its improved variants including, Long Short-Term Memory Networks (LSTMs) and Gated Recurrent Units (GRUs) (Monner and Reggia, 2012). Due to their suitability for processing sequential data, these models are rather useful for analyzing time series data, such as repeated strategic thinking test results, neuroimages, biomaterial values, etc. (Othman et al., 2024). For instance, LSTMs in modelling the decline in memory and executive function by incorporating across time-dependent data for better forecasts on future cognitive status (Zuidersma et al., 2019).
A major advantage of temporal modeling is that a large and diverse number of data sources can be incorporated: structural and functional neuroimaging, genomic and other molecular profiles, CSF biomarkers, and HER (Behrad and Saniee Abadeh, 2022). The holistic approach also allows for understanding cognitive changes and making prognosis concerning transitions from MCI to AD or other further stages of neurodegenerative disorder (Lombardo et al., 2024). For instance, integrating MRI with tau and amyloid beta biomarkers, makes it easier for the model to identify more sensitive changes in the brain’s structure and function (Behrad and Saniee Abadeh, 2022). Personalized temporal modelling extends the prediction mechanism by incorporating subject specific differences. Depression, age, type 2 diabetes, hypertension, smoking, obesity, APOE genotype, education, income, and occupation, are some of the predictors of progressive decline and capturing this data is necessary for accurate prognosis at the individual level (McFall et al., 2019).
Temporal constructors within the clustering techniques can also generate groups of patients sharing similar disease progress for better categorization of patient populations for clinical trials or for targeted treatment (Cui et al., 2023). The temporal modeling remains a problem in the following ways: dealing with gaps in available data, requiring vast amounts of data collected over time, and interpretability of prediction (Liu et al., 2023). Data imputation prevents data loss, while transfer learning applies existing algorithms to a new dataset, improving model effectiveness for clinical uses. This can be achieved while temporal modeling advances and becomes connected with other new constantly monitoring techniques such as wearables that can alert for Cognitive decline at even earlier stages than the current status (Arevalo-Rodriguez et al., 2021). They all have potential for helping decrease the load of the neurodegenerative ailments and enhance individual experiences of patients.
4.3. Using deep learning for MCI vs. AD stage differentiation
Deep learning has been shown to be a helpful method for distinguishing between AD and mild MCI, two critical stages in the continuum of cognitive decline and an area where the differential diagnosis of neurodegenerative diseases has been chronically understudied (Grueso and Viejo-Sobera, 2021). While MCI is a mild cognitive disability that makes people develop dementia over time, AD is a severe cognitive disability that entails severe memory impairment, disability, and general neuronal loss. It is very important to be able to differentiate between these stages correctly so that interventions may be addressed appropriately (López et al., 2020).
CNNs are considered best suited for deep learning models when addressing neuroimaging data including MRI and PET scans (Tufail et al., 2022). These models also exclude hierarchical features from imaging data that reflect the structural and functional brain changes related to MCI and AD. CNNs can spot early atrophy in areas like hippocampus and entorhinal cortex in which moderate AD development occurs (Woźniak et al., 2021). This automated feature extraction helps bypass the preprocessing step required for each scan and greatly upgrade diagnostic thoroughness (Jin and Shi, 1998). The temporal modeling part helps analyze eventual transitions from MCI to AD, which proves the dynamic nature of neurodegeneration (Bilgel et al., 2017). A more advanced differentiation is achieved by multimodal deep learning that combines different data feeder. Thus, adding neuroimaging with other structural and functional data, including genetic profiles (for example, APOE status), cerebrospinal fluid biomarkers (for instance, amyloid beta and tau), and clinical evaluation enhances the coverage of knowledge about the disease (Venugopalan et al., 2021). These multimodal models are typically better than single-modality ones because they combine different type of data.
However, there are challenges when employing deep learning for the differentiation between MCI and AD. A major challenge is the requirement for large annotated data sets to be able to ‘teach’ the models well (Li et al., 2015). Additionally, the “black box” nature of deep learning models can limit their clinical acceptance. In response to this, it is possible to design explainable artificial intelligence (XAI) methods that will enhance the understandability of the model’s prediction to the clinicians (Chaddad et al., 2023). There has also been recent progress in fine-tuning such models to work with small datasets in order to build models that work well with limited resources (Agarwal et al., 2021). With the help of deep learning, the study of MCI and AD can be differentiated at an earlier stage and the focus can be shifted to accurate early diagnosis, as well as identifying more individual approaches to prologue patients with neurodegenerative diseases (Raza and Abbas, 2024).
4.4. Risk stratification and personalized treatment approaches
Risk stratification and targeted management on the other hand have become part of precision medicine as it brings out who among the people has different level of risk for something and how treatment option should be administered to the patient (Reddy et al., 2020). It enables the identification of who among the population is at different extents of disease danger and how treatment should be implemented to the patient.
These treatments are particularly important in chronic diseases including cardiovascular diseases, cancer, diabetes and neurodegenerative diseases (Xu et al., 2021). Risk stratification means placing patients into categories that are expected to develop disease or other adverse outcomes along a specific continuum (Jin et al., 2020). It relies on data obtained from clinical practice, genomic and biomolecular, and patient’s individual parameters and behaviors to produce customer-specific risk models (Jin et al., 2020). This means “risk stratification” that may require blood levels of cholesterol, blood pressure, family history, and in some cases also genetic risks to calculate for example the probability of myocardial infarction or stroke as for cardiovascular disease (Paquette and Baass, 2018).
Machine learning models go further and analyze big data to identify nonlinear patterns in the data that improve the accuracy of risk predictions and allow for the early identification of high-risk individuals (Mariani et al., 2021). Personalized treatment strategies go beyond risk stratification by creating tailored interventions based on a person’s individual characteristics. In oncology, for instance, genetic profiling of tumors enables the use of targeted therapies aimed at certain mutations, including HER2 inhibitors for breast cancer patients with HER2 overexpression (Oh and Bang, 2019).
In psychiatry, for example, we can select antidepressants based on genetic variants that change drug metabolism, a process that should reduce side effects and improve effectiveness (Tansey et al., 2012). The strategies could be potentiated by the integration of wearable technologies and digital health platforms permitting continuous monitoring of physiological parameters towards real-time risk stratification and progressive updating of therapy (Lu et al., 2020). Continuous glucose monitoring systems, for example, offer real-time feedback that can inform personalized dials for insulin dosing for patients with diabetes, reducing the risk for complications such as hypoglycemia (Breton et al., 2018).
Despite the fact that the leverage of risk stratification, and tailored treatments open up large avenues, its business is framed by several major hurdles for ‘big data’-driven research, including data protection, inequality in access to advanced technologies, and the necessity of solid clinical backing for the adoption of predictive models and personal drugs (Parikh et al., 2019). Moreover, for machine learning algorithms to be widely adopted they must be interpretable and create confidence in their outputs (Rudin et al., 2021).
AI and genomics advances are about to take the burden off of personalized medicine delivery challenges (Zahra et al., 2024). Focusing on targeted care in such approaches transforms healthcare from a reactive paradigm that limits disease prevention, therapeutic improvements, and resource allocation, to an anticipative approach that envisages all of these parameters in a managed and systematic manner. So, risk stratification and personalized treatments will be the future of clinical practice, which should ultimately result in improved outcomes both at the patient and the population levels.
4.5. Deep radiomics in PET and MRI imaging for Alzheimer’s disease diagnosis
Deep radiomics is an AI-driven approach that in PET and MRI is revolutionizing AD diagnosis by extracting quantitative biomarkers that provide deeper insights into disease progression. By leveraging deep learning techniques, such as CNNs, radiomics can automatically extract high-dimensional features from imaging data, revealing subtle metabolic and structural abnormalities that may not be visually detectable (Stefano, 2024). The major advantages of deep radiomics in AD diagnosis include early detection of neurodegenerative changes, objective biomarker quantification, and the ability to assess disease progression longitudinally (Ding et al., 2021) Additionally, integrating radiomics with multimodal AI models that incorporate genetics, clinical data, and cognitive assessments can enhance diagnostic accuracy and patient stratification for clinical trials (Seo et al., 2022) However, challenges such as standardizing radiomics pipelines, ensuring data generalizability across diverse populations, and obtaining regulatory approval (e.g., FDA, CE Marking, GDPR compliance) must be addressed for widespread clinical adoption (Obuchowicz et al., 2025). Additionally, deep radiomics combined with explainable AI (xAI) techniques will refine AI-assisted diagnostic pipelines, making them more trustworthy, transparent, and clinically actionable for improving Alzheimer’s disease detection, prognosis, and patient care (Alharthi et al., 2024).
Tables 6 and 7 summarize the comparison of deep radiomics with traditional radiomics and benefits of deep radiomics in neuroimaging, respectively, for AD diagnosis. These AI-driven approaches improve the differentiation between normal aging, MCI, and AD, enhancing the ability to predict cognitive decline and personalize treatment strategies.
Table 6.
Comparison of radiomics and deep learning in neuroimaging.
Aspect | Traditional radiomics | Deep radiomics |
---|---|---|
Feature Extraction | Predefined features (texture, intensity, shape) | Learned hierarchical features via CNNs (Zhou et al., 2021) |
Data Interpretation | Limited to manual analysis | Integrates imaging with clinical/genetic/cognitive data (Shu et al., 2021) |
Predictive Modeling | Relies on classical statistical models | AI-driven prediction of disease progression (Peng et al., 2025) |
Table 7.
Comprehensive overview of deep radiomics in neuroimaging for AD diagnosis.
Modality | Biomarker/Feature | Purpose/Application in AD | Deep Radiomics Benefit |
---|---|---|---|
FDG-PET | Glucose metabolism | Detects hypometabolism in AD-prone regions | Identifies subtle imaging patterns via automated feature extraction (Zhou et al., 2021) |
Amyloid PET | Amyloid-beta | Identifies amyloid plaque accumulation | Enhances early diagnosis by detecting AD before clinical symptoms appear (Rasi et al., 2024) |
Tau PET | Tau protein | Correlates with neurodegeneration and cognitive decline | Assists in predicting MCI to AD conversion (Feng et al., 2024) |
Structural MRI (sMRI) | Brain atrophy | Measures hippocampal and medial temporal lobe atrophy | Detects minute volumetric changes using voxel-based analysis (Shih et al., 2024) |
Functional MRI (fMRI) | Neuronal connectivity | Assesses activity and connectivity changes in AD | Identifies microstructural abnormalities via texture analysis (Leandrou, 2021) |
Diffusion Tensor Imaging (DTI) | White matter integrity | Evaluates microstructural changes in neuronal pathways | Combines MRI with PET for enhanced diagnostic accuracy (Afrazeh, 2024) |
4.6. Explainable AI (xAI) in Alzheimer’s disease diagnosis
With the increasing reliance on deep learning models for medical diagnostics, ensuring trust, transparency, and interpretability is crucial, especially in critical conditions like Alzheimer’s disease (Nazir et al., 2024). Traditional AI models, particularly convolutional neural networks (CNNs) and transformer-based models, often operate as black boxes, making it difficult for clinicians to understand, validate, and trust the decisions made by these models (Lai, 2024). xAI addresses this challenge by providing interpretability techniques that make AI decisions more transparent and reliable for clinical applications (Adeniran et al., 2024).
4.6.1. Key xAI techniques for Alzheimer’s disease diagnosis
Several interpretability methods are applied to deep learning models used in Alzheimer’s disease detection from brain imaging (e.g., MRI, PET scans) and clinical data. Some of the most widely used techniques include:
Grad-CAM (Gradient-weighted Class Activation Mapping), generates heatmaps by identifying the most important regions in an image that influenced the model’s prediction (Song et al., 2024). It calculates the gradient of the predicted class with respect to the last convolutional layer and highlights relevant features in the input image (Asokan and Seshadri, 2023) this technique is used in CNN-based Alzheimer’s classifiers to highlight key regions in MRI or PET scans that contribute to the prediction (e.g., hippocampal atrophy). It also helps radiologists to validate AI predictions and ensures that the model focuses on medically relevant areas rather than noise (Toumaj et al., 2024).
SHAP (Shapley Additive Explanations): SHAP values are derived from game theory to assign an importance score to each feature in a model’s prediction. It explains how each input feature (e.g., age, brain volume, cognitive test scores) contributes to the final decision (Braithwaite et al., 2020). In AD these techniques are used in tabular models (e.g., decision trees, deep learning) to analyze biomarker data, clinical assessments, and genetic risk factors. It facilitates clinicians to understand which features (e.g., APOE4 genotype, CSF biomarkers) are most influential in predicting disease progression (Sarica et al., 2024).
LIME (Local Interpretable Model-agnostic Explanations) perturbs the input data and observes how predictions change, fitting a simpler, interpretable model (e.g., linear regression) to explain the prediction locally (Zafar and Khan, 2021). In AD, it is applied to black-box classifiers to assess why an AI system labeled a brain scan as AD, MCI (Mild Cognitive Impairment), or normal aging. It helps in validating model predictions by ensuring consistency with known clinical insights. In AD, it is applied to black-box classifiers to assess why an AI system labeled a brain scan as AD, MCI or normal aging (Hasan Saif et al., 2024).
Clinical usability is strengthened by leveraging xAI techniques. This is done by improving trust as physicians are more likely to adopt AI-driven diagnostic tools when they can see and verify why a model makes specific predictions. As far as Regulatory compliance is concerned, Explainable models align with healthcare AI regulations (e.g., FDA, GDPR in AI/ML) by making AI decisions auditable and interpretable. Additionally, xAI enhance patient outcomes by making personalized treatment possible as clinicians can understand AI predictions and make informed interventions based on transparent reasoning (Vimbi et al., 2024).
5. Challenges in data standardization and generalization in imaging
Variability in imaging protocols and data quality presents significant challenges in applying imaging technology across different fields, including medical imaging, remote sensing, and scientific imaging. Typically, radiological results are assessed qualitatively by the human eye and expert interpretation (Gao et al., 2019). However, quantitative assessment can provide more valuable insights that are often missed by subjective human evaluation. For personalized treatment, objective assessment becomes crucial for accurate quantification. A major issue is the variability in results, which complicates generalization. This variability arises from several factors (Zeng et al., 2023).
The main challenges of deep learning models in detecting early AD are their interpretability and limitations in clinical adoption due to lack of explainability (Vimbi et al., 2024). Another important thing is that, unlike traditional machine learning models that produce clear decision boundaries, deep neural networks act as complex “black boxes” making it hard to understand the logic behind their predictions (Vimbi et al., 2024). This lack of transparency unsettles clinicians and regulators, because the decision process is opaque, so it is not possible to validate whether models are finding biomarkers related to the disease, or whether they are being influenced by irrelevant correlations, such as scanner artifacts or demographic biases.
Recent advancements in explainable AI (XAI) techniques have been proposed to improve model interpretability to overcome these challenges (Machlev et al., 2022). Techniques like Gradient-weighted Class Activation Mapping (Grad-CAM), SHAP, and LIME provide visualizations of the most salient brain segments that inform model predictions, leading clinicians to better evaluate the reasonableness of AI outputs (S Band et al., 2023). The introduction of attention mechanisms in transformer-based models also show the potential of such an approach to enhance the interpretability of predictions mapping MRI and PET scans directly to disease-relevant biomarkers (Sibilano et al., 2024). Furthermore, hybrid models that combine deep learning with manually crafted radiomic features like hippocampal atrophy serve to unify AI-based findings and known clinical relevance, as both diagnostics can be used together to explain the outputs of deep learning-derived models (Adadi and Berrada, 2018; Xu et al., 2024). These approaches can alleviate resistance towards black-box models, by providing visual or quantitative explanations regarding decision-making processes.
In addition to explainability, uncertainty estimation techniques (e.g., Bayesian deep learning) can help quantify the confidence level for various AI predictions (Nor et al., 2022). This is especially useful for clinical applications, as healthcare professionals can balance AI-generated diagnoses in light of the model’s confidence level (Sai et al., 2024). In addition, interactive AI systems in which clinicians query and refine outputs from the model may be an intuitive way to increase trust and usability (Nasarian et al., 2024). Empirical investigations further support dataset curation and improvement with annotated lesion maps, further directing deep learning models to relevant, biological features and lessening potential dependencies on unrelated features or patterns (Rai et al., 2024). Importantly, benchmark drives within the research community for interpretability will be critical for real-world clinical deployment of future models with the levels of transparency required.
5.1. Diversity in imaging protocols and data quality
The diversity of imaging protocols and data quality can lead to inconsistencies that affect the accuracy, dependability, and precision of the results. Imaging protocols encompass the steps, methods, and settings used to acquire imaging data. Key factors contributing to variability in imaging protocols include (Keenan et al., 2021).
Imaging Type: Different types of imaging techniques can be employed, such as MRI, CT scans, X-rays, ultrasounds, or satellite imagery (Hussain et al., 2022).
Device Parameters: Variations in exposure time, resolution, contrast, and scanning parameters can impact the quality of the obtained images.
Data Processing: The final steps in processing the captured data can influence the outcome.
The variation in imaging protocols is influenced by factors such as the type of equipment used, operator differences, institutional standards, and technological advancements. For instance, different organizations may use different models of MRI machines with distinct hardware, software, and scanning protocols. Even within the same institution, two machines might vary due to differences in magnetic field strength, coil types, and scan sequences. These inconsistencies make it difficult to standardize and compare images across devices (Hussain et al., 2022). Then comes the quantitative imaging biomarkers, these are quantitative figures from pictures that can describe the nature of physiological developments, pathologies or treatment comebacks (Radwan et al., 2024).
The major concepts related to imaging biomarkers are:
Transition to digital imaging
Standardization and control of imaging
Accuracy and precision in results
Reliability and reproducibility (generalizability)
Statistical tools for assessing variability
Variation in MRI and CT scam along with the sources of variation
Computer aided software used in imaging (Wood, 2020).
5.1.1. Human factors and technological advancements
Human factors contribute significantly to variability in imaging results. Differences in operator training, experience, and expertise can lead to divergent approaches to data acquisition and analysis. Additionally, while new technologies generally improve imaging quality, older technologies become outdated, further complicating standardization efforts. Variations in machine settings—such as resolution, scanning time, and slice thickness—also affect image quality (Hussain et al., 2022).
5.1.2. Standardization efforts
To address these challenges, significant efforts have been made to standardize imaging protocols and results across institutions and devices. One such initiative is the Digital Imaging and Communications in Medicine (DICOM) standard, which aims to ensure high-quality imaging and protocol standardization (Hussain et al., 2022). Additionally, advancements in artificial intelligence (AI) and machine learning are being leveraged to automatically adjust imaging factors, optimizing image quality regardless of the equipment used.
5.2. Addressing small sample sizes and imbalanced data
Learning from imbalanced data remains a complex issue in machine learning and data mining. Despite two decades of research on this topic, the challenge persists, especially in binary classification tasks. The advent of big data has provided insights into imbalanced learning, but it has also introduced new challenges (Sandfort et al., 2019). Hybrid approaches—combining data-level and algorithm-level methods—are becoming more prominent in the field. Current approaches focus not only on the division of data but also on the inherent difficulties associated with the nature of the data itself (Krawczyk, 2016). The presence of small cohorts in machine learning is an important consideration which can reduce the reliability, statistical power, and generalization of the model and can cause overfitting that occurs when the model captures noise instead of the original patterns, leading to poor presentation to resolve this issue expansion of dataset is needed while including methods such as random forest to improve quality (Kokol et al., 2022). However, implementing nested cross-validation can reduce biases even with limited data. Moreover, a careful model selection can also prevent overfitting in small sample sizes (Vabalas et al., 2019).
5.3. Generalizability across diverse populations and settings
Inferential goals in research typically involve gaining insights into a particular target population based on study observations. However, the results can vary depending on the chosen population. In some cases, such as randomized controlled trials (RCTs) or policy research, the sample may not truly represent the target population (Zhang et al., 2020b). Clinicians, in particular, may face challenges in making prognostic evaluations and can be prone to inaccuracies in their conclusions (Korevaar et al., 2023). Reproducibility, defined as the ability to generate precise predictions across individuals not part of the study, remains a key challenge (Degtiar and Rose, 2023). Further, working on extensive and balanced datasets will reduce bias and increase the generalizability of findings across different cohorts and adapting self-directed models can help in increasing model performance, which can then detect diseases like chronic obstructive pulmonary disease (COPD) across different ethnic groups (D Almeida et al., 2024). The use of models that are specific to definite settings can enhance the outcomes and reliability, especially in locations where data is frequently available (Yang et al., 2023).
5.4. Handling missing data and noise in multimodal datasets
Multimodal learning, which combines insights from different data sources, is increasingly applied to address real-world challenges, such as noisy or incomplete data (Cao et al., 2022). Despite this, current research lacks a deep investigation into how noise and data incompleteness affect prognostic outcomes in multimodal datasets. Understanding the impact of these factors is essential for improving the accuracy and reliability of predictions derived from multimodal data sources (Degtiar and Rose, 2023).
5.4.1. Challenges in deploying AI models for Alzheimer’s disease diagnosis
While AI-driven models show promise in diagnosing Alzheimer’s disease (AD), several hurdles must be addressed before they can be widely adopted in clinical settings. These include regulatory approval, real-world validation, and seamless integration into existing workflows. Table 8 summarizes that for AI to be effectively deployed in Alzheimer’s disease diagnosis, it must be regulatory-compliant, validated in real-world settings, and seamlessly integrated into clinical workflows. Overcoming these hurdles will ensure that AI serves as a trustworthy, efficient, and clinically useful tool in early detection and patient care.
Table 8.
Key challenges and practical implications of AI models.
Hurdle | Challenges | Potential Solutions |
---|---|---|
Regulatory Approval | AI models must comply with strict healthcare regulations (e.g., FDA, CE Marking, GDPR in AI/ML) (Pantanowitz et al., 2024) | Develop explainable AI (xAI) models to enhance transparency and regulatory acceptance (Gaur and Gaur, 2024) |
Variability in data sources (MRI/PET scans, biomarkers) leads to model inconsistencies (Stefano, 2024) | Conduct standardized multi-center trials to validate model reliability (Kumar et al., 2024) | |
Real-World Validation | AI models trained on research datasets often fail in diverse patient populations (Daneshjou et al., 2021) | Perform external validation on real-world hospital data to assess generalizability (Birkenbihl et al., 2020) |
Lack of longitudinal studies hinders AI’s ability to track disease progression (Habibi et al., 2025) | Integrate AI with longitudinal datasets to improve early detection and monitoring (Kale et al., 2024) | |
Integration into Clinical Workflows | AI tools must seamlessly work with existing Electronic Health Records (EHRs) and PACS (Picture Archiving and Communication Systems) (Kaur and Sharma, 2024) | Develop interoperable AI solutions compatible with major healthcare IT systems (Adegoke et al., 2025) |
Physicians may lack trust in AI predictions due to the black-box nature of deep learning (Lai, 2024) | Incorporate explainable AI techniques (Grad-CAM, SHAP, LIME) to improve clinician confidence (Tuan, 2024) | |
Time constraints in clinical practice limit AI adoption (Adler-Milstein et al., 2022) | Design AI systems with user-friendly interfaces that provide rapid, interpretable insights (Li et al., 2024) |
5.5. Future directions (these lines needs to be added in existing conclusion)
To enhance Alzheimer’s AI models, future research should focus on, combining multiple xAI techniques for better interpretability, integrating multimodal data (imaging + genetics + clinical features) for a comprehensive diagnosis and developing interactive AI interfaces that allow doctors to query and adjust AI-driven predictions.
6. Conclusion
The integration of deep learning with multimodal brain imaging marks a significant step forward in the early diagnosis and treatment of Alzheimer’s disease. By leveraging AI’s power to analyze complex datasets, researchers can achieve more accurate, timely, and personalized approaches to AD diagnostics and treatment. However, several challenges remain, particularly regarding data standardization, model generalization, and clinical applicability. Future efforts should focus on improving the interpretability of AI models and overcoming technical and ethical barriers to facilitate their broader adoption in clinical settings.
Funding Statement
The author(s) declare that no financial support was received for the research and/or publication of this article.
Author contributions
MR: Conceptualization, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing. SH: Validation, Visualization, Writing – original draft, Writing – review & editing. SJ: Validation, Visualization, Writing – original draft, Writing – review & editing. NH: Visualization, Writing – original draft, Writing – review & editing. KB: Validation, Visualization, Writing – original draft, Writing – review & editing. SW: Validation, Visualization, Writing – original draft, Writing – review & editing. MA: Visualization, Writing – review & editing.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that Gen AI was used in the creation of this manuscript. In preparing this work, the author utilized ChatGPT 3.5 to enhance the clarity and readability of the English text. After employing the ChatGPT 3.5, the author carefully reviewed and edited the content as necessary. The author assumes full accountability for the publication’s content.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
- Abbas M. K., Raza M. L., Zaidi S. S. H., Khan B. M., Heinemann U. (2019). Seizure forecasting using single robust linear feature as correlation vector of seizure-like events in brain slices preparation in vitro. Neurol. Res. 41, 99–109. doi: 10.1080/01616412.2018.1532481, PMID: [DOI] [PubMed] [Google Scholar]
- Aberathne I., Kulasiri D., Samarasinghe S. (2023). Detection of Alzheimer’s disease onset using MRI and PET neuroimaging: longitudinal data analysis and machine learning. Neural Regen. Res. 18, 2134–2140. doi: 10.4103/1673-5374.367840, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Abrol A., Calhoun V., Plis S., Salman M., Silva R., Du Y., et al. (2021). Deep learning encodes robust discriminative neuroimaging representations to outperform standard machine learning. Nat. Commun. 12:353. doi: 10.1038/s41467-020-20655-6, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Adadi A., Berrada M. (2018). Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE Access 6, 52138–52160. doi: 10.1109/ACCESS.2018.2870052 [DOI] [Google Scholar]
- Adegoke K., Adegoke A., Dawodu D., Bayowa A., Adekoya A. (2025). Interoperability in digital healthcare: Enhancing consumer health and transforming care systems.
- Adeniran A. A., Onebunne A. P., William P. (2024). Explainable AI (XAI) in healthcare: enhancing trust and transparency in critical decision-making. World J. Adv. Res. Rev 23, 2647–2658. doi: 10.30574/wjarr.2024.23.3.2936 [DOI] [Google Scholar]
- Adler-Milstein J., Aggarwal N., Ahmed M., Castner J., Evans B. J., Gonzalez A. A., et al. (2022). Meeting the moment: addressing barriers and facilitating clinical adoption of artificial intelligence in medical diagnosis. NAM Perspect. 2022:10.31478/202209c. doi: 10.31478/202209c, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Afrazeh F. (2024). Advances in imaging analysis for understanding brain disorders. Int. J. Sci. Appl. Res. 4, 1–08. doi: 10.54756/IJSAR.2024.15, PMID: 40071581 [DOI] [Google Scholar]
- Agarwal D., Marques G., Díez I., Martín M., Zapirain B., Martín-Rodríguez F. (2021). Transfer learning for Alzheimer’s disease through neuroimaging biomarkers: A systematic review. Sensors (Basel, Switzerland) 21:7259. doi: 10.3390/s21217259, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Agosta F., Pievani M., Geroldi C., Copetti M., Frisoni G. B., Filippi M. (2012). Resting state fMRI in Alzheimer's disease: beyond the default mode network. Neurobiol. Aging 33, 1564–1578. doi: 10.1016/j.neurobiolaging.2011.06.007, PMID: [DOI] [PubMed] [Google Scholar]
- Ahmed S., Choi K. Y., Lee J. J., Kim B. C., Kwon G.-R., Lee K. H., et al. (2019). Ensembles of patch-based classifiers for diagnosis of Alzheimer diseases. IEEE Access 7, 73373–73383. doi: 10.1109/ACCESS.2019.2920011, PMID: 39573497 [DOI] [Google Scholar]
- Airlangga G. (2024). Advancing Alzheimer’s diagnosis: a comparative analysis of deep learning architectures on multidimensional health data. J. Inform. Ekonomi Bisnis 6, 810–814. doi: 10.37034/infeb.v6i4.1046 [DOI] [Google Scholar]
- Albajes-Eizagirre A., Solanes A., Fullana M., Ioannidis J., Fusar-Poli P., Torrent C., et al. (2019). Meta-analysis of voxel-based neuroimaging studies using seed-based d mapping with permutation of subject images (SDM-PSI). J. Vis. Exp. 153:e59841. doi: 10.3791/59841 [DOI] [PubMed] [Google Scholar]
- Alharthi A., Alqurashi A., Alharbi T., Alammar M., Aldosari N., Bouchekara H., et al. (2024). The role of explainable AI in revolutionizing human health monitoring. Ar xiv preprint ar xiv: 2409.07347.
- Alowais S. A., Alghamdi S. S., Alsuhebany N., Alqahtani T., Alshaya A. I., Almohareb S. N., et al. (2023). Revolutionizing healthcare: the role of artificial intelligence in clinical practice. BMC Med. Educ. 23:689. doi: 10.1186/s12909-023-04698-z, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alshamlan H., Alwassel A., Banafa A., Alsaleem L. (2024). Improving Alzheimer's disease prediction with different machine learning approaches and feature selection techniques. Diagnostics (Basel, Switzerland) 14:2237. doi: 10.3390/diagnostics14192237, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alsubaie M. G., Luo S., Shaukat K. (2024). Alzheimer’s disease detection using deep learning on neuroimaging: a systematic review. Mach. Learn. Knowl. Extract. 6, 464–505. doi: 10.3390/make6010024, PMID: 40053772 [DOI] [Google Scholar]
- Amira B. R., Faouzi B., Hamid A. (2015). “Medical application: diagnosis of Alzheimer disease from MRI and documents.” in 2015 7th International Conference on Modelling, Identification and Control (ICMIC). pp. 1–6.
- Aramadaka S., Mannam R., Sankara Narayanan R., Bansal A., Yanamaladoddi V. R., Sarvepalli S. S., et al. (2023). Neuroimaging in Alzheimer's disease for early diagnosis: a comprehensive review. Cureus 15:e38544. doi: 10.7759/cureus.38544, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Arevalo-Rodriguez I., Smailagic N., Roqué-Figuls M., Ciapponi A., Sanchez-Perez E., Giannakou A., et al. (2021). Mini-mental state examination (MMSE) for the early detection of dementia in people with mild cognitive impairment (MCI). Cochrane Database Syst. Rev. 7:Cd010783. doi: 10.1002/14651858.CD010783.pub3, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Asaduzzaman M., Alom M. K., Karim M. E. (2025). ALZENET: deep learning-based early prediction of Alzheimer's disease through magnetic resonance imaging analysis. Telemat. Inform. Rep. 17:100189. doi: 10.1016/j.teler.2025.100189 [DOI] [Google Scholar]
- Asokan S., Seshadri A. (2023). Hierarchical spatial feature-CNN employing grad-CAM for enhanced segmentation and classification in Alzheimer's and Parkinson's disease diagnosis via MRI. Traitement du Signal 40, 2769–2778. doi: 10.18280/ts.400637 [DOI] [Google Scholar]
- Atri A. (2019). The Alzheimer's disease clinical Spectrum: diagnosis and management. Med. Clin. North Am. 103, 263–293. doi: 10.1016/j.mcna.2018.10.009, PMID: [DOI] [PubMed] [Google Scholar]
- Babb de Villiers C., Kroese M., Moorthie S. (2020). Understanding polygenic models, their development and the potential application of polygenic scores in healthcare. J. Med. Genet. 57, 725–732. doi: 10.1136/jmedgenet-2019-106763, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Band S., Yarahmadi A., Hsu C.-C., Biyari M., Sookhak M., Ameri R., et al. (2023). Application of explainable artificial intelligence in medical health: a systematic review of interpretability methods. Inform. Med. Unlocked 40:101286. doi: 10.1016/j.imu.2023.101286 [DOI] [Google Scholar]
- Behrad F., Saniee Abadeh M. (2022). An overview of deep learning methods for multimodal medical data mining. Expert Syst. Appl. 200:117006. doi: 10.1016/j.eswa.2022.117006 [DOI] [Google Scholar]
- Bilgel M., Koscik R., An Y., Prince J., Resnick S., Johnson S., et al. (2017). Temporal order of Alzheimer's disease-related cognitive marker changes in BLSA and WRAP longitudinal studies. J. Alzheimers Dis. 59, 1335–1347. doi: 10.3233/JAD-170448, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Binnewijzend M. A., Schoonheim M. M., Sanz-Arigita E., Wink A. M., van der Flier W. M., Tolboom N., et al. (2012). Resting-state fMRI changes in Alzheimer's disease and mild cognitive impairment. Neurobiol. Aging 33, 2018–2028. doi: 10.1016/j.neurobiolaging.2011.07.003, PMID: [DOI] [PubMed] [Google Scholar]
- Birkenbihl C., Emon M. A., Vrooman H., Westwood S., Lovestone S., Consortium A., et al. (2020). Differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia-lessons for translation into clinical practice. EPMA J. 11, 367–376. doi: 10.1007/s13167-020-00216-z, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bouts M. J., Möller C., Hafkemeijer A., van Swieten J. C., Dopper E., van der Flier W. M., et al. (2018). Single subject classification of Alzheimer’s disease and behavioral variant frontotemporal dementia using anatomical, diffusion tensor, and resting-state functional magnetic resonance imaging. J. Alzheimers Dis. 62, 1827–1839. doi: 10.3233/JAD-170893, PMID: [DOI] [PubMed] [Google Scholar]
- Braithwaite B., Paananen J., Taipale H., Tanskanen A., Tiihonen J., Hartikainen S., et al. (2020). Detection of medications associated with Alzheimer's disease using ensemble methods and cooperative game theory. Int. J. Med. Inform. 141:104142. doi: 10.1016/j.ijmedinf.2020.104142, PMID: [DOI] [PubMed] [Google Scholar]
- Breton M., Patek S., Lv D., Schertz E., Robic J., Pinnata J., et al. (2018). Continuous glucose monitoring and insulin informed advisory system with automated titration and dosing of insulin reduces glucose variability in type 1 diabetes mellitus. Diabetes Technol. Ther. 20, 531–540. doi: 10.1089/dia.2018.0079, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cao Y., Fu L., Wu J., Peng Q., Nie Q., Zhang J., et al. (2022). Integrated analysis of multimodal single-cell data with structural similarity. Nucleic Acids Res. 50:e121. doi: 10.1093/nar/gkac781, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Castellano G., Esposito A., Lella E., Montanaro G., Vessio G. (2024). Automated detection of Alzheimer’s disease: a multi-modal approach with 3D MRI and amyloid PET. Sci. Rep. 14:5210. doi: 10.1038/s41598-024-56001-9, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chaddad A., Peng J., Xu J., Bouridane A. (2023). Survey of explainable AI techniques in healthcare. Sensors (Basel, Switzerland) 23:634. doi: 10.3390/s23020634, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Choudhury C., Goel T., Tanveer M. (2024). A coupled-GAN architecture to fuse MRI and PET image features for multi-stage classification of Alzheimer’s disease. Inform. Fusion 109:102415. doi: 10.1016/j.inffus.2024.102415 [DOI] [Google Scholar]
- Chouliaras L., O’Brien J. T. (2023). The use of neuroimaging techniques in the early and differential diagnosis of dementia. Mol. Psychiatry 28, 4084–4097. doi: 10.1038/s41380-023-02215-8, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cui C., Li X., Zhang C., Guan W., Wang M. (2023). Temporal-relational hypergraph tri-attention networks for stock trend prediction. Pattern Recogn. 143:109759. doi: 10.1016/j.patcog.2023.109759 [DOI] [Google Scholar]
- Almeida D., Norajitra T., Lüth C. T., Wald T., Weru V., Nolden M., et al. (2024). How do deep-learning models generalize across populations? Cross-ethnicity generalization of COPD detection. Insights Imaging 15:198. doi: 10.1186/s13244-024-01781-x, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dachena C., Casu S., Lodi M. B., Fanti A., Mazzarella G. (2020). “Application of MRI, fMRI and cognitive data for Alzheimer’s disease detection.” in Paper presented at the 2020 14th European conference on antennas and propagation (EuCAP).
- Dakdareh S. G., Abbasian K. (2024). Diagnosis of Alzheimer’s disease and mild cognitive impairment using convolutional neural networks. J. Alzheimers Dis. Rep. 8, 317–328. doi: 10.3233/ADR-230118, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Damoiseaux J. S. (2012). Resting-state fMRI as a biomarker for Alzheimer's disease? Alzheimers Res. Ther. 4, 8–2. doi: 10.1186/alzrt106, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Daneshjou R., Smith M. P., Sun M. D., Rotemberg V., Zou J. (2021). Lack of transparency and potential bias in artificial intelligence data sets and algorithms: a scoping review. JAMA Dermatol. 157, 1362–1369. doi: 10.1001/jamadermatol.2021.3129, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Degtiar I., Rose S. (2023). A review of generalizability and transportability. Ann. Rev. Stat. Appl. 10, 501–524. doi: 10.1146/annurev-statistics-042522-103837 [DOI] [Google Scholar]
- Ding Y., Sohn J., Kawczynski M., Trivedi H., Harnish R., Jenkins N., et al. (2019). A deep learning model to predict a diagnosis of Alzheimer disease by using 18F-FDG PET of the brain. Radiology 290, 456–464. doi: 10.1148/radiol.2018180958, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ding Y., Zhao K., Che T., Du K., Sun H., Liu S., et al. (2021). Quantitative radiomic features as new biomarkers for Alzheimer’s disease: An amyloid PET study. Cereb. Cortex 31, 3950–3961. doi: 10.1093/cercor/bhab061, PMID: [DOI] [PubMed] [Google Scholar]
- Du L., Langhough R. E., Wilson R. E., Reyes R. E. R., Hermann B. P., Jonaitis E. M., et al. (2024a). Longitudinal plasma phosphorylated-tau 217 and other related biomarkers in a non-demented Alzheimer's risk-enhanced sample. Alzheimers Dement. 20, 6183–6204. doi: 10.1002/alz.14100, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Du L., Roy S., Wang P., Li Z., Qiu X., Zhang Y., et al. (2024b). Unveiling the future: advancements in MRI imaging for neurodegenerative disorders. Ageing Res. Rev. 95:102230. doi: 10.1016/j.arr.2024.102230, PMID: [DOI] [PubMed] [Google Scholar]
- Du Z., Yang Y., Zheng J., Li Q., Lin D., Li Y., et al. (2020). Accurate prediction of coronary heart disease for patients with hypertension from electronic health records with big data and machine-learning methods: model development and performance evaluation. JMIR Med. Inform. 8:e17257. doi: 10.2196/17257, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Echle A., Rindtorff N. T., Brinker T. J., Luedde T., Pearson A. T., Kather J. N. (2021). Deep learning in cancer pathology: a new generation of clinical biomarkers. Br. J. Cancer 124, 686–696. doi: 10.1038/s41416-020-01122-x, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Faruqui S. H. A., Du Y., Meka R., Alaeddini A., Li C., Shirinkam S., et al. (2019). Development of a deep learning model for dynamic forecasting of blood glucose level for type 2 diabetes mellitus: secondary analysis of a randomized controlled trial. JMIR Mhealth Uhealth 7:e14452. doi: 10.2196/14452, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Feng J., Huang Y., Zhang X., Yang Q., Guo Y., Xia Y., et al. (2024). Research and application progress of radiomics in neurodegenerative diseases. Meta-Radiol. 2:100068. doi: 10.1016/j.metrad.2024.100068, PMID: 40094570 [DOI] [Google Scholar]
- Feng W., Halm-Lutterodt N. V., Tang H., Mecum A., Mesregah M. K., Ma Y., et al. (2020). Automated MRI-based deep learning model for detection of Alzheimer's disease process. Int. J. Neural Syst. 30:2050032. doi: 10.1142/S012906572050032X, PMID: [DOI] [PubMed] [Google Scholar]
- Galić I., Habijan M., Leventić H., Romić K. (2023). Machine learning empowering personalized medicine: a comprehensive review of medical image analysis methods. Electronics 12:4411. doi: 10.3390/electronics12214411 [DOI] [Google Scholar]
- Gao Y., Liu Y., Wang Y., Shi Z., Yu J. (2019). A universal intensity standardization method based on a many-to-one weak-paired cycle generative adversarial network for magnetic resonance images. IEEE Trans. Med. Imaging 38, 2059–2069. doi: 10.1109/TMI.2019.2894692, PMID: [DOI] [PubMed] [Google Scholar]
- Gaur L., Gaur D. (2024). “Explainable Artificial Intelligence (XAI)” in On Neurogenerative diseases AI and neuro-degenerative diseases: Insights and solutions (Cham, UK: Springer; ), 63–72. [Google Scholar]
- Grueso S., Viejo-Sobera R. (2021). Machine learning methods for predicting progression from mild cognitive impairment to Alzheimer’s disease dementia: a systematic review. Alzheimers Res. Ther. 13:162. doi: 10.1186/s13195-021-00900-w, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Guan J. (2019). Artificial intelligence in healthcare and medicine: promises, ethical challenges and governance. Chin. Med. Sci. J. 34, 76–83. doi: 10.24920/003611, PMID: [DOI] [PubMed] [Google Scholar]
- Guleria H. V., Luqmani A. M., Kothari H., Phukan P., Patil S., Pareek P., et al. (2023). Enhancing the breast histopathology image analysis for Cancer detection using variational autoencoder. Int. J. Environ. Res. Public Health 20:4244. doi: 10.3390/ijerph20054244, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Habeck C., Stern Y., Initiative T. (2010). Multivariate data analysis for neuroimaging data: overview and application to Alzheimer’s disease. Cell Biochem. Biophys. 58, 53–67. doi: 10.1007/s12013-010-9093-0, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Habibi F., Ghaderkhani S., Shokoohi M., Banari T., Morsali M., Abadi R. N. S., et al. (2025). Harnessing artificial intelligence in Alzheimer's disease management: navigating ethical challenges in AI. AI Ethics, 1–17. doi: 10.1007/s43681-025-00673-0 [DOI] [Google Scholar]
- Hasan Saif F., Al-Andoli M. N., Bejuri W. M. Y. W. (2024). Explainable AI for Alzheimer detection: a review of current methods and applications. Appl. Sci. 14:10121. doi: 10.3390/app142210121 [DOI] [Google Scholar]
- Hofmann-Apitius M., Ball G., Gebel S., Bagewadi S., De Bono B., Schneider R., et al. (2015). Bioinformatics mining and modeling methods for the identification of disease mechanisms in neurodegenerative disorders. Int. J. Mol. Sci. 16, 29179–29206. doi: 10.3390/ijms161226148, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hojjati S. H., Ebrahimzadeh A., Khazaee A., Babajani-Feremi A., Initiative A. (2018). Predicting conversion from MCI to AD by integrating rs-fMRI and structural MRI. Comput. Biol. Med. 102, 30–39. doi: 10.1016/j.compbiomed.2018.09.004, PMID: [DOI] [PubMed] [Google Scholar]
- Hu M., Ang K.-K., Zhang H., Nardi C. (2023). Applications of deep learning to neurodevelopment in pediatric imaging: achievements and challenges. Appl. Sci. 13:2302. doi: 10.3390/app13042302 [DOI] [Google Scholar]
- Huang Y., Xu J., Zhou Y., Tong T., Zhuang X., Initiative A. (2019). Diagnosis of Alzheimer’s disease via multi-modality 3D convolutional neural network. Front. Neurosci. 13:509. doi: 10.3389/fnins.2019.00509, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hussain S., Mubeen I., Ullah N., Shah S. S. U. D., Khan B. A., Zahoor M., et al. (2022). Modern diagnostic imaging technique applications and risk factors in the medical field: a review. Bio. Med. Res. Int. 2022:5164970. doi: 10.1155/2022/5164970, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ibrahim B., Suppiah S., Ibrahim N., Mohamad M., Hassan H. A., Nasser N. S., et al. (2021). Diagnostic power of resting-state fMRI for detection of network connectivity in Alzheimer's disease and mild cognitive impairment: a systematic review. Hum. Brain Mapp. 42, 2941–2968. doi: 10.1002/hbm.25369, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jellinger K. A. (2022). Recent update on the heterogeneity of the Alzheimer’s disease spectrum. J. Neural Transm. 129, 1–24. doi: 10.1007/s00702-021-02449-2, PMID: [DOI] [PubMed] [Google Scholar]
- Jin G., Lv J., Yang M., Wang M., Zhu M., Wang T., et al. (2020). Genetic risk, incident gastric cancer, and healthy lifestyle: a meta-analysis of genome-wide association studies and prospective cohort study. Lancet Oncol. 21, 1378–1386. doi: 10.1016/S1470-2045(20)30460-5, PMID: [DOI] [PubMed] [Google Scholar]
- Jin J., Shi J. (1998). Automatic feature extraction of waveform signals for in-process diagnostic performance improvement. J. Intell. Manuf. 12, 257–268. [Google Scholar]
- Jo T., Nho K., Saykin A. (2019). Deep learning in Alzheimer's disease: diagnostic classification and prognostic prediction using neuroimaging data. Front. Aging Neurosci. 11:220. doi: 10.3389/fnagi.2019.00220, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Johnson K. A., Fox N. C., Sperling R. A., Klunk W. E. (2012). Brain imaging in Alzheimer disease. Cold Spring Harb. Perspect. Med. 2:a006213. doi: 10.1101/cshperspect.a006213, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kale M. B., Wankhede N. L., Pawar R. S., Ballal S., Kumawat R., Goswami M., et al. (2024). AI-driven innovations in Alzheimer's disease: integrating early diagnosis, personalized treatment, and prognostic modelling. Ageing Res. Rev. 101:102497. doi: 10.1016/j.arr.2024.102497, PMID: [DOI] [PubMed] [Google Scholar]
- Kaur G., Sharma S. (2024). Innovative Technologies for Healthcare Service Productivity Innovative Technologies for increasing service productivity. Hershey, Pennsylvania, USA: IGI Global, 274–292. [Google Scholar]
- Keenan K., Delfino J., Jordanova K., Poorman M., Chirra P., Chaudhari A., et al. (2021). Challenges in ensuring the generalizability of image quantitation methods for MRI. Med. Phys. 48, 4681–4691. doi: 10.1002/mp.15195 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Khan H., Saddam. (2024). A novel feature map enhancement technique integrating residual CNN and transformer for Alzheimer diseases diagnosis. ArXiv 2405:12986. doi: 10.48550/arXiv.2405.12986 [DOI] [Google Scholar]
- Khan T., Alkon D. (2015). Alzheimer's disease cerebrospinal fluid and neuroimaging biomarkers: diagnostic accuracy and relationship to drug efficacy. J. Alzheimers Dis. 46, 817–836. doi: 10.3233/JAD-150238, PMID: [DOI] [PubMed] [Google Scholar]
- Khatun M., Islam M. M., Rahman Rifat H., Bin Shahid M. S., Talukder M. A., Uddin M. A. (2023). “Hybridized convolutional neural networks and Long short-term memory for improved Alzheimer’s disease diagnosis from MRI scans.” in 2023 26th International Conference on Computer and Information Technology, ICCIT 2023.
- Kokol P., Kokol M., Zagoranski S. (2022). Machine learning on small size samples: a synthetic knowledge synthesis. Sci. Prog. 105:368504211029777. doi: 10.1177/00368504211029777, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Korevaar S., Tennakoon R., Bab-Hadiashar A. (2023). Failure to achieve domain invariance with domain generalization algorithms: an analysis in medical imaging. IEEE Access 11, 39351–39372. doi: 10.1109/ACCESS.2023.3268704 [DOI] [Google Scholar]
- Krawczyk B. (2016). Learning from imbalanced data: open challenges and future directions. Prog. Artif. Intellig. 5, 221–232. doi: 10.1007/s13748-016-0094-0 [DOI] [Google Scholar]
- Kumar R., Waisberg E., Ong J., Paladugu P., Amiri D., Saintyl J., et al. (2024). Artificial intelligence-based methodologies for early diagnostic precision and personalized therapeutic strategies in neuro-ophthalmic and neurodegenerative pathologies. Brain Sci. 14:1266. doi: 10.3390/brainsci14121266, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kunst J., Marecek R., Klobusiakova P., Balazova Z., Anderkova L., Nemcova-Elfmarkova N., et al. (2019). Patterns of Grey matter atrophy at different stages of Parkinson's and Alzheimer's diseases and relation to cognition. Brain Topogr. 32, 142–160. doi: 10.1007/s10548-018-0675-2, PMID: [DOI] [PubMed] [Google Scholar]
- Kuznetsov V., Lee H. K., Maurer-Stroh S., Molnár M. J., Pongor S., Eisenhaber B., et al. (2013). How bioinformatics influences health informatics: usage of biomolecular sequences, expression profiles and automated microscopic image analyses for clinical needs and public health. Health Inform. Sci. Syst. 1:2. doi: 10.1186/2047-2501-1-2, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lai T. (2024). Interpretable medical imagery diagnosis with self-attentive transformers: a review of explainable AI for health care. Bio. Med. Inform. 4, 113–126. doi: 10.3390/biomedinformatics4010008, PMID: 40053772 [DOI] [Google Scholar]
- Leandrou S. (2021). Quantitative texture analysis in MR imaging in the assessment of Alzheimer’s disease. City, University of London.
- Lee J., Renslo J., Wong K., Clifford T. G., Beutler B. D., Kim P. E., et al. (2024). Current trends and applications of PET/MRI hybrid imaging in neurodegenerative diseases and Normal aging. Diagnostics 14:585. doi: 10.3390/diagnostics14060585, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li F., Tran L., Thung K. H., Ji S., Shen D., Li J. (2015). A robust deep model for improved classification of AD/MCI patients. IEEE J. Biomed. Health Inform. 19, 1610–1616. doi: 10.1109/JBHI.2015.2429556, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li H., Zech J., Hong D., Ghamisi P., Schultz M., Zipf A. (2022). Leveraging open street map and multimodal remote sensing data with joint deep learning for wastewater treatment plants detection. Int. J. Appl. Earth Observ. Geoinform. 110:102804. doi: 10.1016/j.jag.2022.102804, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li X., Zheng H., Chen J., Zong Y., Yu L. (2024). User interaction interface design and innovation based on artificial intelligence technology. J. Theory Pract. Eng. Sci. 4, 1–8. doi: 10.53469/jtpes.2024.04(03).01 [DOI] [Google Scholar]
- Li Y., Qian B., Zhang X., Liu H. (2020). Graph neural network-based diagnosis prediction. Big Data 8, 379–390. doi: 10.1089/big.2020.0070, PMID: [DOI] [PubMed] [Google Scholar]
- Lin W., Gao Q., Du M., Chen W., Tong T. (2021). Multiclass diagnosis of stages of Alzheimer's disease using linear discriminant analysis scoring for multimodal data. Comput. Biol. Med. 134:104478. doi: 10.1016/j.compbiomed.2021.104478, PMID: [DOI] [PubMed] [Google Scholar]
- Liu Q., Zhao X., Wang G. (2023). A clustering ensemble method for cell type detection by multiobjective particle optimization. IEEE/ACM Trans. Comput. Biol. Bioinform. 20, 1–14. doi: 10.1109/TCBB.2021.3132400, PMID: [DOI] [PubMed] [Google Scholar]
- Logan R., Ferreira Da Silva M., Ganguly A., Miller S. J., Schcolnicov N., Indani A., et al. (2021). Deep convolutional neural networks with ensemble learning and generative adversarial networks for Alzheimer’s disease image data classification. Front. Aging Neurosci. 13:720226. doi: 10.3389/fnagi.2021.720226, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lombardo F. L., Lorenzini P., Mayer F., Massari M., Piscopo P., Bacigalupo I., et al. (2024). Development of a prediction model of conversion to Alzheimer's disease in people with mild cognitive impairment: the statistical analysis plan of the INTERCEPTOR project. Diagn. Progn. Res. 8:11. doi: 10.1186/s41512-024-00172-6, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- López M., Turrero A., Cuesta P., Rodríguez-Rojo I., Barabash A., Marcos A., et al. (2020). A multivariate model of time to conversion from mild cognitive impairment to Alzheimer’s disease. Gero Sci. 42, 1715–1732. doi: 10.1007/s11357-020-00260-7, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Loreto F., Verdi S., Kia S. M., Duvnjak A., Hakeem H., Fitzgerald A., et al. (2024). Alzheimer's disease heterogeneity revealed by neuroanatomical normative modeling. Alzheimers Dement. 16:e12559. doi: 10.1002/dad2.12559, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lu L., Zhang J., Xie Y., Gao F., Xu S., Wu X., et al. (2020). Wearable health devices in health care: narrative systematic review. JMIR Mhealth Uhealth 8:e18907. doi: 10.2196/18907, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Machlev R., Heistrene L., Perl M., Levy K. Y., Belikov J., Mannor S., et al. (2022). Explainable artificial intelligence (XAI) techniques for energy and power systems: review, challenges and opportunities. Energy AI 9:100169. doi: 10.1016/j.egyai.2022.100169 [DOI] [Google Scholar]
- Madan S., Lentzen M., Brandt J., Rueckert D., Hofmann-Apitius M., Fröhlich H. (2024). “Transformer models in biomedicine” in BMC medical informatics and decision making, vol. 24 (Bio Med Central Ltd.) doi: 10.1186/s12911-024-02600-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Majee A., Gupta A., Raha S., Das S. (2024). Enhancing MRI-based classification of Alzheimer’s disease with explainable 3D hybrid compact convolutional transformers.
- Mann M., Kumar C., Zeng W. F., Strauss M. T. (2021). Artificial intelligence for proteomics and biomarker discovery. Cell Syst. 12, 759–770. doi: 10.1016/j.cels.2021.06.006, PMID: [DOI] [PubMed] [Google Scholar]
- Mao S., Sejdic E. (2023). A review of recurrent neural network-based methods in computational physiology. IEEE Trans. Neural Netw. Learn. Syst. 34, 6983–7003. doi: 10.1109/TNNLS.2022.3145365, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mariani M., Biney F., Tweneboah O. (2021). Analyzing medical data by using statistical learning models. Mathematics 9:968. doi: 10.3390/math9090968 [DOI] [Google Scholar]
- McFall G. P., McDermott K. L., Dixon R. A. (2019). Modifiable risk factors discriminate memory trajectories in non-demented aging: precision factors and targets for promoting healthier brain aging and preventing dementia. J. Alzheimers Dis. 70, S101–s118. doi: 10.3233/JAD-180571, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Miller C., Portlock T., Nyaga D. M., O'Sullivan J. M. (2024). A review of model evaluation metrics for machine learning in genetics and genomics. Front. Bioinform. 4:1457619. doi: 10.3389/fbinf.2024.1457619, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mirkin S., Albensi B. (2023). Should artificial intelligence be used in conjunction with neuroimaging in the diagnosis of Alzheimer’s disease? Front. Aging Neurosci. 15:10944233. doi: 10.3389/fnagi.2023.1094233, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Monner D., Reggia J. A. (2012). A generalized LSTM-like training algorithm for second-order recurrent neural networks. Neur. Netw. 25, 70–83. doi: 10.1016/j.neunet.2011.07.003, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mosconi L., De Santi S., Rusinek H., Convit A., De Leon M. J. (2004). Magnetic resonance and PET studies in the early diagnosis of Alzheimer’s disease. Expert. Rev. Neurother. 4, 831–849. doi: 10.1586/14737175.4.5.831, PMID: [DOI] [PubMed] [Google Scholar]
- Mueller S. G., Weiner M. W., Thal L. J., Petersen R. C., Jack C., Jagust W., et al. (2005). The Alzheimer's disease neuroimaging initiative. Neuroimaging Clin. N. Am. 15, 869–877. doi: 10.1016/j.nic.2005.09.008, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nasarian E., Alizadehsani R., Acharya U. R., Tsui K.-L. (2024). Designing interpretable ML system to enhance trust in healthcare: a systematic review to proposed responsible clinician-AI-collaboration framework. Inform. Fusion 108:102412. doi: 10.1016/j.inffus.2024.102412 [DOI] [Google Scholar]
- Nazir A., Assad A., Hussain A., Singh M. (2024). Alzheimer’s disease diagnosis using deep learning techniques: datasets, challenges, research gaps and future directions. Int. J. Syst. Assur. Eng. Manag., 1–35. doi: 10.1007/s13198-024-02441-5, PMID: 40094045 [DOI] [Google Scholar]
- Nor A. K. M., Pedapati S. R., Muhammad M., Leiva V. (2022). Abnormality detection and failure prediction using explainable Bayesian deep learning. Methodol. Case Study Indust. Data 10:155. doi: 10.3390/math10040554, PMID: 40053772 [DOI] [Google Scholar]
- Obuchowicz R., Lasek J., Wodziński M., Piórkowski A., Strzelecki M., Nurzynska K. (2025). Artificial intelligence-empowered radiology—current status and critical review. Diagnostics 15:282. doi: 10.3390/diagnostics15030282, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Oh D., Bang Y. (2019). HER2-targeted therapies — a role beyond breast cancer. Nat. Rev. Clin. Oncol. 17, 33–48. doi: 10.1038/s41571-019-0268-3, PMID: [DOI] [PubMed] [Google Scholar]
- Al Olaimat M., Bozdag S. (2024). TA-RNN: an attention-based time-aware recurrent neural network architecture for electronic health records. Bioinformatics 40, i169–i179. doi: 10.1093/bioinformatics/btae264, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Othman G. B., Ynineb A. R., Yumuk E., Farbakhsh H., Muresan C., Birs I. R., et al. (2024). Artificial intelligence-driven prognosis of respiratory mechanics: forecasting tissue Hysteresivity using Long short-term memory and continuous sensor data. Sensors (Basel) 24:5544. doi: 10.3390/s24175544, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pantanowitz L., Hanna M., Pantanowitz J., Lennerz J., Henricks W. H., Shen P., et al. (2024). Regulatory aspects of AI-ML. Mod. Pathol. 37:100609. doi: 10.1016/j.modpat.2024.100609 [DOI] [PubMed] [Google Scholar]
- Paproki A., Salvado O., Fookes C. (2024). Synthetic data for deep learning in Computer Vision & Medical Imaging: a means to reduce data bias. ACM Comput. Surv. 56:37. doi: 10.1145/3663759 [DOI] [Google Scholar]
- Paquette M., Baass A. (2018). Predicting cardiovascular disease in familial hypercholesterolemia. Curr Opin Lipidol. 29, 299–306. doi: 10.1097/MOL.0000000000000519 [DOI] [PubMed] [Google Scholar]
- Parihar A., Swami P. D. (2024). Analysis of EEG signals with the use of wavelet transform for accurate classification of Alzheimer disease, frontotemporal dementia and healthy subjects using machine learning models. Fusion Pract. Appl. 14. doi: 10.54216/FPA.140203 [DOI] [Google Scholar]
- Parikh R., Gdowski A., Patt D., Hertler A., Mermel C., Bekelman J. (2019). Using big data and predictive analytics to determine patient risk in oncology. Am. Soc. Clin. Oncol. Educ. Book 39, e53–e58. doi: 10.1200/EDBK_238891, PMID: [DOI] [PubMed] [Google Scholar]
- Park C., Joo G., Roh M., Shin S., Yum S., Yeo N. Y., et al. (2024). Predicting the progression of mild cognitive impairment to Alzheimer’s dementia using recurrent neural networks with a series of neuropsychological tests. J. Clin. Neurol. 20, 478–486. doi: 10.3988/jcn.2023.0289, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Parmar H., Nutter B., Long R., Antani S., Mitra S. (2020). Spatiotemporal feature extraction and classification of Alzheimer’s disease using deep learning 3D-CNN for fMRI data. J. Med. Imaging 7:6001. doi: 10.1117/1.JMI.7.5.056001, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Peng D., Huang W., Liu R., Zhong W. (2025). From pixels to prognosis: radiomics and AI in Alzheimer’s disease management. Front. Neurol. 16:1536463. doi: 10.3389/fneur.2025.1536463, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Peng J., Zhu X., Wang Y., An L., Shen D. (2019). Structured sparsity regularized multiple kernel learning for Alzheimer’s disease diagnosis. Pattern Recogn. 88, 370–382. doi: 10.1016/j.patcog.2018.11.027, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pirrone D., Weitschek E., Di Paolo P., De Salvo S., De Cola M. (2022). EEG signal processing and supervised machine learning to early diagnose Alzheimer’s disease. Appl. Sci. 12:5413. doi: 10.3390/app12115413 [DOI] [Google Scholar]
- Polikar R., Tilley C., Hillis B., Clark C. M. (2010). “Multimodal EEG, MRI and PET data fusion for Alzheimer's disease diagnosis.” in Paper presented at the 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology. [DOI] [PubMed]
- Porsteinsson A. P., Isaacson R. S., Knox S., Sabbagh M. N., Rubino I. (2021). Diagnosis of early Alzheimer's disease: clinical practice in 2021. J. Prev. Alzheimers Dis. 8, 371–386. doi: 10.14283/jpad.2021.23, PMID: [DOI] [PubMed] [Google Scholar]
- Rabeh A. B., Benzarti F., Amiri H. (2023). “CNN-SVM for prediction of Alzheimer disease in early step.” in 2023 International Conference on Control, Automation and Diagnosis (ICCAD). pp. 1–6.
- Radwan A., Osman I., Shehata M. (2024). Universal medical imaging model for domain generalization with data privacy. ArXiv abs/2407.14719. doi: 10.48550/arXiv.2407.14719 [DOI] [Google Scholar]
- Rai H. M., Yoo J., Razaque A. (2024). Comparative analysis of machine learning and deep learning models for improved cancer detection: a comprehensive review of recent advancements in diagnostic techniques. Expert Syst. Appl. 255:124838. doi: 10.1016/j.eswa.2024.124838 [DOI] [Google Scholar]
- Raja R., Rosenberg G. A., Caprihan A. (2018). MRI measurements of blood-brain barrier function in dementia: a review of recent studies. Neuropharmacology 134, 259–271. doi: 10.1016/j.neuropharm.2017.10.034, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rao G., Gao H., Wang X., Zhang J., Ye M., Rao L. (2023). MRI measurements of brain hippocampus volume in relation to mild cognitive impairment and Alzheimer disease: a systematic review and meta-analysis. Medicine 102:e34997. doi: 10.1097/MD.0000000000034997, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rasi R., Guvenis A., Initiative A. S. D. N. (2024). Predicting amyloid positivity from FDG-PET images using radiomics: a parsimonious model. Comput. Methods Prog. Biomed. 247:108098. doi: 10.1016/j.cmpb.2024.108098, PMID: [DOI] [PubMed] [Google Scholar]
- Raza M. L., Imam M. H., Zehra W., Jamil S. (2024a). Neuro-inflammatory pathways in COVID-19-induced central nervous system injury: implications for prevention and treatment strategies. Exp. Neurol. 382:114984. doi: 10.1016/j.expneurol.2024.114984, PMID: [DOI] [PubMed] [Google Scholar]
- Raza M. L., Bhojani A., Batool S. K., Zehra D. (2024b). Non pharmacological approaches for neurodegenerative diseases: a narrative review. Exp. Gerontol. 198:112620. doi: 10.1016/j.exger.2024.112620, PMID: [DOI] [PubMed] [Google Scholar]
- Raza M. L., Abbas M. K. (2024). “Artificial intelligence in drug discovery and repurposing for neurodegenerative disorders” in Emerging Frontiers in translation neurodegeneration. Ed. 1st (USA: CRC Press, Taylor & Francis; ), 79–104. [Google Scholar]
- Reddy S., Allan S., Coghlan S., Cooper P. (2020). A governance model for the application of AI in health care. J. Am. Med. Inform. Assoc. 27, 491–497. doi: 10.1093/jamia/ocz192, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Romano J. D., Truong V., Kumar R., Venkatesan M., Graham B. E., Hao Y., et al. (2024). The Alzheimer’s Knowledge Base: a knowledge graph for Alzheimer disease research. J. Med. Internet Res. 26:e46777. doi: 10.2196/46777, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rudin C., Chen C., Chen Z., Huang H., Semenova L., Zhong C. (2021). Interpretable machine learning: fundamental principles and 10 grand challenges. ArXiv abs/2103.11251. doi: 10.48550/arXiv.2103.11251 [DOI] [Google Scholar]
- Sai S., Gaur A., Sai R., Chamola V., Guizani M., Rodrigues J. J. P. C. (2024). Generative AI for transformative healthcare: a comprehensive study of emerging models, applications, case studies, and limitations. IEEE Access 12, 31078–31106. doi: 10.1109/ACCESS.2024.3367715 [DOI] [Google Scholar]
- Saleem T. J., Zahra S. R., Wu F., Alwakeel A., Alwakeel M., Jeribi F., et al. (2022). Deep learning-based diagnosis of Alzheimer’s disease. J. Person. Med. 12:815. doi: 10.3390/jpm12050815 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sandfort V., Yan K., Pickhardt P., Summers R. (2019). Data augmentation using generative adversarial networks (cycle GAN) to improve generalizability in CT segmentation tasks. Sci. Rep. 9:16884. doi: 10.1038/s41598-019-52737-x, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sarica A., Pelagi A., Aracri F., Arcuri F., Quattrone A., Quattrone A., et al. (2024). Sex differences in conversion risk from mild cognitive impairment to Alzheimer’s disease: an explainable machine learning study with random survival forests and SHAP. Brain Sci. 14:201. doi: 10.3390/brainsci14030201, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Seo Y., Jang H., Lee H. (2022). Potential applications of artificial intelligence in clinical trials for Alzheimer’s disease. Life 12:275. doi: 10.3390/life12020275, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shah S. M. A. H., Khan M. Q., Rizwan A., Jan S. U., Samee N. A., Jamjoom M. M. (2024). Computer-aided diagnosis of Alzheimer’s disease and neurocognitive disorders with multimodal bi-vision transformer (BiViT). Pattern. Anal. Applic. 27. doi: 10.1007/s10044-024-01297-6, PMID: 40094045 [DOI] [Google Scholar]
- Sharma H., Zerbe N., Klempert I., Hellwich O., Hufnagl P. (2017). Deep convolutional neural networks for automatic classification of gastric carcinoma using whole slide images in digital histopathology. Comp. Med. Imaging Graph. 61, 2–13. doi: 10.1016/j.compmedimag.2017.06.001, PMID: [DOI] [PubMed] [Google Scholar]
- Shih D.-H., Wu Y.-H., Wu T.-W., Wang Y.-K., Shih M.-H. (2024). Classifying dementia severity using MRI Radiomics analysis of the Hippocampus and machine learning. IEEE Access 12, 160030–160051. doi: 10.1109/ACCESS.2024.3483833 [DOI] [Google Scholar]
- Shu H. (2016). High dimensional dependent data analysis for neuroimaging (doctoral dissertation).
- Shu Z.-Y., Mao D.-W., Xu Y.-Y., Shao Y., Pang P.-P., Gong X.-Y. (2021). Prediction of the progression from mild cognitive impairment to Alzheimer’s disease using a radiomics-integrated model. Ther. Adv. Neurol. Disord. 14:17562864211029551. doi: 10.1177/17562864211029551, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shukla A., Tiwari R., Tiwari S. (2023). Alzheimer’s disease detection from fused PET and MRI modalities using an ensemble classifier. Mach. Learn. Knowl. Extract. 5, 512–538. doi: 10.3390/make5020031 [DOI] [Google Scholar]
- Sibilano E., Buongiorno D., Lassi M., Grippo A., Bessi V., Sorbi S., et al. (2024). Understanding the role of self-attention in a transformer model for the discrimination of SCD from MCI using resting-state EEG. IEEE J. Biomed. Health Inform. 28, 3422–3433. doi: 10.1109/JBHI.2024.3390606, PMID: [DOI] [PubMed] [Google Scholar]
- Song B., Yoshida S., Initiative A. S. D. N. (2024). Explainability of three-dimensional convolutional neural networks for functional magnetic resonance imaging of Alzheimer’s disease classification based on gradient-weighted class activation mapping. PLoS One 19:e0303278. doi: 10.1371/journal.pone.0303278, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Song J., Zheng J., Li P., Lu X., Zhu G., Shen P. (2021). An effective multimodal image fusion method using MRI and PET for Alzheimer's disease diagnosis. Front. Digit. Health 3:637386. doi: 10.3389/fdgth.2021.637386, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stefano A. (2024). Challenges and limitations in applying radiomics to PET imaging: possible opportunities and avenues for research. Comput. Biol. Med. 179:108827. doi: 10.1016/j.compbiomed.2024.108827, PMID: [DOI] [PubMed] [Google Scholar]
- Suma K., Raghavan D., Ganesh P. (2022). “Deep learning for Alzheimer's disease detection using multimodal MRI-PET fusion.” in Paper presented at the 2022 4th International Conference on Circuits, Control, Communication and Computing (I4C).
- Suppiah S., Didier M.-A., Vinjamuri S. (2019). The who, when, why, and how of PET amyloid imaging in management of Alzheimer’s disease—review of literature and interesting images. Diagnostics 9:65. doi: 10.3390/diagnostics9020065, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tang C., Wei M., Sun J., Wang S., Zhang Y. (2023). CsAGP: detecting Alzheimer’s disease from multimodal images via dual-transformer with cross-attention and graph pooling. J. King Saud Univ. Comp. Inform. Sci. 35:101618. doi: 10.1016/j.jksuci.2023.101618, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tansey K., Guipponi M., Perroud N., Bondolfi G., Domenici E., Evans D., et al. (2012). Genetic predictors of response to serotonergic and noradrenergic antidepressants in major depressive disorder: a genome-wide analysis of individual-level data and a meta-analysis. PLoS Med. 9:e1001326. doi: 10.1371/journal.pmed.1001326, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Toepper M., Beblo T., Thomas C., Driessen M. (2008). Early detection of Alzheimer's disease: a new working memory paradigm. Int. J. Geriatr. Psychiatry 23, 272–278. doi: 10.1002/gps.1873, PMID: [DOI] [PubMed] [Google Scholar]
- Tong L., Shi W., Isgut M., Zhong Y., Lais P., Gloster L., et al. (2024). Integrating multi-omics data with EHR for precision medicine using advanced artificial intelligence. IEEE Rev. Biomed. Eng. 17, 80–97. doi: 10.1109/RBME.2023.3324264, PMID: [DOI] [PubMed] [Google Scholar]
- Toumaj S., Heidari A., Shahhosseini R., Jafari Navimipour N. (2024). Applications of deep learning in Alzheimer’s disease: a systematic literature review of current trends, methodologies, challenges, innovations, and future directions. Artif. Intell. Rev. 58:44. doi: 10.1007/s10462-024-11041-5 [DOI] [Google Scholar]
- Tuan D. A. (2024). Bridging the gap between black box AI and clinical practice: Advancing explainable AI for trust, ethics, and personalized healthcare diagnostics.
- Tufail A. B., Anwar N., Othman M. T. B., Ullah I., Khan R. A., Yong K., et al. (2022). Early-stage Alzheimer’s disease categorization using PET neuroimaging modality and convolutional neural networks in the 2D and 3D domains. Sensors (Basel, Switzerland) 22:4609. doi: 10.3390/s22124609, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Vabalas A., Gowen E., Poliakoff E., Casson A. J. (2019). Machine learning algorithm validation with a limited sample size. PLoS One 14:e0224365. doi: 10.1371/journal.pone.0224365, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Vadapalli S., Abdelhalim H., Zeeshan S., Ahmed Z. (2022). Artificial intelligence and machine learning approaches using gene expression and variant data for personalized medicine. Brief. Bioinform. 23:191. doi: 10.1093/bib/bbac191, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Vashishtha A., Acharya A. K., Swain S. (2023). Hybrid model: deep learning method for early detection of Alzheimer’s disease from MRI images. Biomed. Pharmacol. J. 16, 1617–1630. doi: 10.13005/bpj/2739 [DOI] [Google Scholar]
- Venugopalan J., Tong L., Hassanzadeh H. R., Wang M. D. (2021). Multimodal deep learning models for early detection of Alzheimer’s disease stage. Sci. Rep. 11:3254. doi: 10.1038/s41598-020-74399-w [DOI] [PMC free article] [PubMed] [Google Scholar]
- Verdi S., Rutherford S., Fraza C., Tosun D., Altmann A., Rakêt L., et al. (2024). Personalizing progressive changes to brain structure in Alzheimer's disease using normative modeling. Alzheimers Dement. 20, 6998–7012. doi: 10.1002/alz.14174, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Vimbi V., Shaffi N., Mahmud M. (2024). Interpreting artificial intelligence models: a systematic review on the application of LIME and SHAP in Alzheimer’s disease detection. Brain Inform. 11:10. doi: 10.1186/s40708-024-00222-1, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang J., Wen S., Liu W., Meng X., Jiao Z. (2024). Deep joint learning diagnosis of Alzheimer's disease based on multimodal feature fusion. Bio Data Mining 17:48. doi: 10.1186/s13040-024-00395-9, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang M., Shao W., Huang S., Zhang D. (2023). Hypergraph-regularized multimodal learning by graph diffusion for imaging genetics based alzheimer’s disease diagnosis. Med. Image Anal. 89:102883. doi: 10.1016/j.media.2023.102883, PMID: [DOI] [PubMed] [Google Scholar]
- Wang Y., Liu S., Spiteri A. G., Huynh A. L. H., Chu C., Masters C. L., et al. (2024). Understanding machine learning applications in dementia research and clinical practice: a review for biomedical scientists and clinicians. Alzheimers Res. Ther. 16:175. doi: 10.1186/s13195-024-01540-6, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Warren J. D., Fletcher P. D., Golden H. L. (2012). The paradox of syndromic diversity in Alzheimer disease. Nat. Rev. Neurol. 8, 451–464. doi: 10.1038/nrneurol.2012.135, PMID: [DOI] [PubMed] [Google Scholar]
- Weinstein H., Scheltens P., Hijdra A., Royen E. (1993). Neuro-imaging in the diagnosis of Alzheimer's disease II. Positron and single photon emission tomography. Clin. Neurol. Neurosurg. 95, 81–91. doi: 10.1016/0303-8467(93)90001-W, PMID: [DOI] [PubMed] [Google Scholar]
- Wen J., Thibeau-Sutre E., Samper-González J., Routier A., Bottani S., Durrleman S., et al. (2020). Convolutional neural networks for classification of Alzheimer's disease: overview and reproducible evaluation. Med. Image Anal. 63:101694. doi: 10.1016/j.media.2020.101694, PMID: [DOI] [PubMed] [Google Scholar]
- Wood M. (2020). Variability and standardization of quantitative imaging. Investig. Radiol. 55, 617–618. doi: 10.1097/RLI.0000000000000667, PMID: [DOI] [PubMed] [Google Scholar]
- Woźniak M., Siłka J., Wieczorek M. (2021). Deep neural network correlation learning mechanism for CT brain tumor detection. Neural Comput. & Applic. 35, 14611–14626. doi: 10.1007/s00521-021-05841-x [DOI] [Google Scholar]
- Wyatt L. S., van Karnenbeek L. M., Wijkhuizen M., Geldof F., Dashtbozorg B. (2024). Explainable artificial intelligence (XAI) for oncological ultrasound image analysis: a systematic. Review 14:8108. doi: 10.3390/app14188108, PMID: 40053772 [DOI] [Google Scholar]
- Xu H., Schwander K., Brown M., Wang W., Waken R., Boerwinkle E., et al. (2021). Lifestyle risk score: handling missingness of individual lifestyle components in meta-analysis of gene-by-lifestyle interactions. Eur. J. Hum. Genet. 29, 839–850. doi: 10.1038/s41431-021-00808-x, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Xu X., Li J., Zhu Z., Zhao L., Wang H., Song C., et al. (2024). A comprehensive review on synergy of multi-modal data and AI technologies in medical diagnosis. Bioengineering (Basel) 11:219. doi: 10.3390/bioengineering11030219, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yang H. S., Pan W., Wang Y., Zaydman M. A., Spies N. C., Zhao Z., et al. (2023). Generalizability of a machine learning model for improving utilization of parathyroid hormone-related peptide testing across multiple clinical centers. Clin. Chem. 69, 1260–1269. doi: 10.1093/clinchem/hvad141, PMID: [DOI] [PubMed] [Google Scholar]
- Zafar M. R., Khan N. (2021). Deterministic local interpretable model-agnostic explanations for stable explainability. Mach. Learn. Knowl. Extract. 3, 525–541. doi: 10.3390/make3030027 [DOI] [Google Scholar]
- Zahra M. A., Al-Taher A., Alquhaidan M., Hussain T., Ismail I., Raya I., et al. (2024). The synergy of artificial intelligence and personalized medicine for the enhanced diagnosis, treatment, and prevention of disease. Drug Metab. Person. Therapy 39, 47–58. doi: 10.1515/dmpt-2024-0003 [DOI] [PubMed] [Google Scholar]
- Zeng J., Zhou J., Liu T. (2023). Robust multimodal sentiment analysis via tag encoding of uncertain missing modalities. IEEE Trans. Multimed. 25, 6301–6314. doi: 10.1109/TMM.2022.3207572 [DOI] [Google Scholar]
- Zhang D., Shen D. (2012). Multi-modal multi-task learning for joint prediction of multiple regression and classification variables in Alzheimer's disease. NeuroImage 59, 895–907. doi: 10.1016/j.neuroimage.2011.09.069, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang L., Wang M., Liu M., Zhang D. (2020a). A survey on deep learning for neuroimaging-based brain disorder analysis. Front. Neurosci. 14:779. doi: 10.3389/fnins.2020.00779, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang L., Wang X., Yang D., Sanford T., Harmon S., Turkbey B., et al. (2020b). Generalizing deep learning for medical image segmentation to unseen domains via deep stacked transformation. IEEE Trans. Med. Imaging 39, 2531–2540. doi: 10.1109/TMI.2020.2973595, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang Y.-D., Dong Z., Wang S.-H., Yu X., Yao X., Zhou Q., et al. (2020). Advances in multimodal data fusion in neuroimaging: overview, challenges, and novel orientation. Inform. Fusion 64, 149–187. doi: 10.1016/j.inffus.2020.07.006, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhao Z., Yeoh P. S. Q., Zuo X., Chuah J. H., Chow C. O., Wu X., et al. (2024). Vision transformer-equipped convolutional neural networks for automated Alzheimer's disease diagnosis using 3D MRI scans. Front. Neurol. 15:1490829. doi: 10.3389/fneur.2024.1490829, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhou J., Zhao M., Yang Z., Chen L., Liu X. (2024). Exploring the value of MRI measurement of hippocampal volume for predicting the occurrence and progression of Alzheimer's disease based on artificial intelligence deep learning technology and evidence-based medicine Meta-analysis. J. Alzheimers Dis. 1275–1288. doi: 10.3233/JAD-230733 [DOI] [PubMed] [Google Scholar]
- Zhou P., Zeng R., Yu L., Feng Y., Chen C., Li F., et al. (2021). Deep-learning radiomics for discrimination conversion of Alzheimer's disease in patients with mild cognitive impairment: a study based on 18F-FDG PET imaging. Front. Aging Neurosci. 13:764872. doi: 10.3389/fnagi.2021.764872, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zuidersma M., Sjöberg L., Pantzar A., Fratiglioni L., Wang H. X. (2019). A bi-factor model of the Montgomery Åsberg depression rating scale and future cognitive impairments in older adults: a 6-year follow-up study. J. Psychiatr. Res. 109, 1–9. doi: 10.1016/j.jpsychires.2018.11.010, PMID: [DOI] [PubMed] [Google Scholar]