Abstract
Background
The Mini-Mental State Examination (MMSE) remains widely used for cognitive screening, yet its performance varies substantially across educational backgrounds. Linear education corrections fail to capture the nonlinear interference patterns among subitems.
Objective
This study aimed to analyze how educational level shapes MMSE subitem contributions and to develop an education-adaptive optimization strategy using support vector machine–based weighting.
Methods
MMSE data from 812 participants were stratified into 4 education groups. Subitem deletion experiments quantified each subitem’s contribution (Δ). Education-specific support vector machine models were then constructed to derive dynamic weighting coefficients. Performance improvements were assessed before and after weighting.
Results
The illiterate group relied heavily on spatial orientation and memory, whereas university-educated individuals depended more on executive and calculation functions. Several education-dependent interference items were identified (eg, visuospatial construction in the primary group and basic orientation tasks in the university group). Dynamic weighting improved accuracy in all cohorts, most notably among illiterate individuals (Δ=7.25%; P=.06), followed by the primary school group (Δ=3.12%; P=.03).
Conclusions
Education-stratified weighting enhances the fairness and interpretability of MMSE-based screening. External validation confirmed generalizability, although multicenter studies are needed.
Introduction
The Mini-Mental State Examination (MMSE) is one of the most widely used cognitive screening tools in clinical and community settings, designed for the rapid detection of cognitive impairment [1]. Since its development by Folstein et al in 1975 [2], the MMSE has become a core instrument for assessing five cognitive domains—orientation, immediate recall, attention or calculation, language, and visuospatial ability—and can typically be completed within 10 to 15 minutes [3]. Its broad adoption reflects its operational simplicity and solid psychometric performance, with reported sensitivity of 80% to 85% and specificity of 75% to 80% for dementia screening [4].
Alzheimer disease is characterized by progressive neurodegeneration, including amyloid-β deposition and structural decline. As Alzheimer disease progresses in a relatively predictable neuroanatomical sequence, MMSE subdomains provide clinically meaningful stage markers: orientation and memory deficits often correspond to early hippocampal involvement, while language and visuospatial impairments reflect later temporoparietal degeneration [5,6]. This multidimensional structure allows the MMSE to map aspects of the disease trajectory beyond a single total score.
Nevertheless, the diagnostic performance of the MMSE varies considerably across educational levels [7,8]. Individuals with higher education often exhibit false-negative results due to two well-recognized mechanisms: compensatory neuroplasticity that delays the manifestation of cognitive symptoms [9] and the use of test-taking strategies that allow them to maintain near-normal scores despite underlying pathology [10,11]. In contrast, individuals with limited formal education show substantially higher false-positive rates, particularly on education-dependent items, such as literacy tasks and object naming, with error rates increasing by approximately 30% to 35% compared with education-neutral instruments [12-15]. As a result, the MMSE demonstrates reduced sensitivity (68%‐72%) and specificity (65%‐70%) when a conventional cutoff score of 24 points is applied across heterogeneous educational backgrounds [16-18]. Current clinical guidelines therefore emphasize the need to incorporate educational history into cognitive assessment to minimize diagnostic inaccuracies [19,20].
Machine learning (ML) has emerged as a transformative approach in medical diagnostics, with particular relevance to cognitive assessment optimization. Support Vector Machines (SVMs) offer several advantages for psychometric refinement, including (1) robust performance in limited-sample settings (typically n<1000) [21,22], (2) effective handling and interpretability of high-dimensional feature spaces [23,24], and (3) strong compatibility with multimodal data integration frameworks that combine neuropsychological assessments with imaging- or biomarker-derived measures [25-28]. Earlier ML classifiers trained solely on MMSE total scores or subitems—most commonly logistic regression, SVM, or random forest—have demonstrated moderate diagnostic performance, with reported accuracies of 72% to 85% and area under the curve (AUC) values of 0.75 to 0.85 [29]. In contrast, multimodal models integrating magnetic resonance imaging (MRI) radiomics, positron emission tomography (PET) signatures, or speech biomarkers consistently achieve accuracies exceeding 90% to 95% [30,31]. This performance gap highlights the need for methodological innovations that enhance unimodal MMSE-based classifiers while preserving their scalability and low implementation cost.
This study introduces an education-sensitive adaptation of the MMSE based on an SVM-guided dynamic weighting framework. We analyzed cognitive screening data (n=812) collected from Chinese tertiary hospitals and community health centers and stratified participants into four educational cohorts: illiterate (0 y), primary (≤6 y), secondary (7‐12 y), and tertiary (≥13 y). The model uses systematic item-response analysis combined with SVM-guided weighting to adjust the relative contribution of each MMSE subitem in an education-specific manner.
This work provides 2 methodological innovations. First, it moves beyond conventional linear education corrections by using nonlinear SVM modeling to identify “cognitive interference” patterns unique to each education level. Second, it introduces a dynamic weighting strategy specifically designed to mitigate the disproportionate false-positive burden experienced by low-literacy populations. Together, these innovations aim to deliver a more equitable and clinically practical adaptation of the MMSE, particularly relevant for resource-limited settings.
Although unimodal or condensed classifiers derived solely from MMSE subitems are highly accessible, they are inherently constrained by limited feature richness and therefore tend to exhibit moderate predictive performance. In contrast, multimodal diagnostic systems—such as those leveraging MRI, PET, radiomics, or speech biomarkers—achieve substantially higher accuracies but remain impractical for widespread screening due to cost, technical requirements, and limited availability in primary care. Thus, methodological advances that enhance the diagnostic utility of unimodal MMSE-based approaches, while maintaining their affordability and scalability, are urgently needed. This study addresses this gap by proposing an education-stratified dynamic weighting method designed to improve predictive performance within the intrinsic limitations of unimodal cognitive classifiers.
Methods
Ethical Considerations
This study followed the principles of the Declaration of Helsinki and was approved by the Ethics Committee of the First Affiliated Hospital of the University of Science and Technology of China (approval number 2024-RE-431). All clinical and community datasets were obtained with authorization from the original data custodians. During data processing, only anonymized records were used, and no personally identifiable information (eg, names, addresses) was accessed, ensuring compliance with the Declaration of Helsinki.
Sample
Cognitive assessment data were obtained from two sources: the Rehabilitation Medicine Department of a tertiary hospital and a community-based cognitive screening program. A total of 812 valid records were included. Among these, 404 samples were collected from outpatients and inpatients with clinically diagnosed cognitive impairment. The inclusion and exclusion criteria are summarized in Table 1.
Table 1. Data inclusion and exclusion criteria.
| Source | Number (n) | Inclusion criteria | Exclusion criteria | Data assessment requirements |
|---|---|---|---|---|
| Clinical dataset | 404 |
|
|
|
| Community dataset | 408 |
|
|
|
MMSE: Mini-Mental State Examination.
Scale Selection and Scoring Criteria
The MMSE was used as the primary cognitive assessment tool. The scale comprises five cognitive domains: orientation, memory, attention and calculation, language, and visuospatial construction. Education-adjusted cutoff scores were based on Chinese normative standards [3]: illiterate group: ≤17, primary school group: ≤20, secondary school group: ≤22, and university group: ≤23.
Data Processing and Grouping
Overview
A total of 1000 assessments were collected initially. After quality control, 188 (18.8%) records were excluded due to missing data (n=107, 10.7%), nonstandardized test administration (n=35, 3.5%), or unqualified assessors (n=46, 4.6%). Before merging the hospital and community datasets, baseline demographic comparability was examined using an independent-samples t test (age) and chi-square test (gender) to minimize potential selection bias.
The final dataset (N=812) was stratified by educational attainment as follows: illiterate (n=108, 13.3%), primary school (n=105, 12.9%), secondary school (n=364, 44.8%), and university (n=235, 28.9%). Within each group, patient and control counts were recorded as illiterate (patients: n=60, 7.4%; and controls: n=48, 5.9%), primary (patients: n=53, 6.5%; and controls: n=52, 6.4%), secondary (patients: n=185, 22.8%; and controls: n=179, 22.0%), and university (patients: n=106, 13.1%; and controls: n=129, 15.9%). This preprocessing ensured a standardized data foundation for subsequent ML procedures.
Model Selection
A SVM classifier with a Radial Basis Function kernel was selected due to its suitability for medium-sized datasets and its capacity to model nonlinear decision boundaries. A stratified 5-fold cross-validation scheme was used to maximize data utilization while preserving the original case-control ratio within each educational group.
Hyperparameter Optimization
Hyperparameters were optimized through a grid search embedded within the cross-validation framework. The search grid included penalty parameter C: {0.1, 1, 5, 10, 100} and kernel coefficient γ: {0.001, 0.01, 0.1, 1}.
For each parameter pair, the model was trained on 4 folds and evaluated on the remaining fold, and the average accuracy across all 5 folds was used to identify the optimal configuration. Education-specific models were then retrained on the full dataset using the selected hyperparameters to ensure maximal predictive performance while avoiding information leakage.
Experimental Design
The overall experimental workflow is presented in Figure 1.
Figure 1. Contribution of individual Mini-Mental State Examination (MMSE)j items across educational strata. Feature importance was quantified using the percentage change in cross-validated accuracy (Δi) after deleting each subitem. We operationally defined critical factors as items whose removal caused a performance decrease of Δ<−1.0 and interference factors as items whose removal increased accuracy by Δ>+0.5. SVM: support vector machine.
The process began with data fusion and initial grouping: clinical and community data were merged to construct the initial dataset (n=812), and participants were stratified into 4 educational groups (ie, illiterate, primary school, secondary school, and university) to ensure subsequent analyses reflected education-specific cognitive characteristics. Next, a Support Vector Machine classifier with a radial basis function kernel was trained using a stratified 5-fold cross-validation scheme, preserving the original case-control ratio within each group. To quantify the contribution of each MMSE subitem, a systematic ablation procedure was performed. For each educational group, each subitem was removed in turn, and an SVM model was retrained. The feature contribution was quantified using the cross-validated change in prediction accuracy (Δ), calculated as:
A negative Δ indicates a positive contribution (performance dropped when the item was removed), while a positive Δ suggests the item acted as noise. Finally, based on the subitem contribution profiles, an education-specific dynamic weighting scheme was constructed to adjust the relative importance of MMSE subitems before classification.
Other Statistical Techniques
Routine statistical analyses were performed using SPSS version 24.0 (IBM Corp), and ML model construction and evaluation were conducted using Python (version 3.10.10; Python Software Foundation)
For categorical variables, chi-square or Fisher exact tests were applied depending on expected frequencies. Continuous variables with a normal distribution were expressed as mean (SD) and compared using independent-samples t tests. Variables not following a normal distribution were summarized using median (IQR) and analyzed using nonparametric tests.
A 2-tailed P<.05 was considered statistically significant.
Results
Demographic Comparison
Following data cleansing, demographic characteristics of the clinical and community datasets were compared to verify baseline homogeneity before integration. As presented in Table 2, differences in age and gender distributions were not statistically significant (P>.05).
Table 2. Demographic characteristics and data composition stratified by educational background.
| Characteristic | Illiterate (n=108) | Primary school (n=105) | Secondary school (n=364) | University (n=235) | P value |
|---|---|---|---|---|---|
| Age, mean (SD) | 59.02 (8.5) | 58.55 (7.8) | 58.62 (8.1) | 56.7 (7.4) | .05 |
| Gender, n (%) | .37 | ||||
| Male | 67 (62.0) | 60 (57.1) | 219 (60.2) | 140 (59.6) | |
| Female | 41 (38.0) | 45 (42.9) | 145 (39.8) | 95 (40.4) | |
| Data source, n (%) | .12 | ||||
| Clinicala (patient) | 60 (55.6) | 53 (50.5) | 185 (50.8) | 106 (45.1) | |
| Communityb (control) | 48 (44.4) | 52 (49.5) | 179 (49.2) | 129 (54.9) |
Hospital dataset: patients diagnosed or evaluated in the tertiary hospital.
Community dataset: participants recruited from community-based cognitive screening programs.
Item-Wise Contribution Analysis
To assess the contribution of individual MMSE items across educational groups, we conducted an item deletion experiment. The change in cross-validated accuracy (Δi) after removing each subitem was used to quantify its importance. A negative Δi (< -1.0) indicated a critical item whose removal impaired performance, while a positive Δi (> +0.5) indicated an interference item whose removal improved accuracy. The results of this analysis are visualized in Figure 2.
Figure 2. Heatmap visualization of feature contribution analysis. The color spectrum reflects prediction accuracy after item deletion. Cooler colors (blue and green) denote "critical features" (accuracy drop), while warmer colors (yellow and red) denote "neutral/interfering features" (accuracy stable/increase). The y-axis represents educational groups with the following sample sizes (N_patient/N_control): illiterate (60/48), primary school (53/52), secondary school (185/179), and university (106/129). DM: delayed memory; EA: execution ability; IM: immediate memory; PP: penmanship and praxis; SS: spatial structure.

Illiterate Group
In the illiterate cohort (baseline accuracy=78.83%), all subitems demonstrated positive contributions. The most influential items were spatial orientation (Δ=−6.58% postremoval), immediate memory (Δ=−3.85%), and writing (Δ=−2.86%). No interference items were identified.
Primary School Group
In the primary school group (baseline accuracy=85.71%), spatial construction functioned as an interference factor (Δ =+0.95%), while calculation showed a negligible effect (Δ=0). Removal of immediate memory, delayed recall, reading, or writing reduced accuracy to 83.81% (Δ=−1.90%).
Secondary School Group
In the secondary school group (baseline accuracy=69.78%), multiple interference factors were observed: spatial orientation (Δ=+3.57%), delayed recall (Δ=+3.02%), executive function (Δ=+2.19%), reading (Δ=+1.64%), immediate memory (Δ=+2.47%), writing (Δ=+0.20%), and spatial construction (Δ=+1.10%). Only temporal orientation (Δ=−0.83%) and calculation ability (Δ=−0.80%) demonstrated positive contributions.
University Group
In the university group (baseline accuracy=71.49%), five interference items were identified: spatial orientation (Δ=+2.98%), calculation (Δ=+2.98%), immediate memory (Δ=+1.70%), naming (Δ=+1.70%), and delayed recall (Δ=+1.42%). Repetition (Δ=−0.57%) and reading (Δ=−0.30%) showed small positive contributions.
Item Weight Analysis
On the basis of these contribution patterns, education-specific weighted coefficients were calculated for all MMSE subitems. The weighted diagnostic score was defined as follows:
where S is the final weighted diagnostic score, vi is the raw score of the i-th MMSE subitem, and wi is the dynamic weighting coefficient assigned to the i-th subitem.
Classification was determined by comparing S to a group-specific classification threshold (T): S<T indicated cognitive impairment, and S≥T indicated normal cognition.
After normalization, the maximum possible weighted score was 60. Education-specific diagnostic thresholds were defined as follows: illiterate (t=30), primary (t=31), secondary (t=32), and university (t=33). As presented in Tables3 4, accuracy improved across all educational strata after weight calibration.
Table 3. Model performance and weighting coefficients for core cognitive domains by educational group.
| Group | Prediction accuracy before weighting | Prediction accuracy after weighting | Orientation (time) | Orientation (space) | Memory (immediate) | Memory (delayed) | Attention and calculation |
|---|---|---|---|---|---|---|---|
| Illiterate (%) | 79.83 | 86.08 | 1 | 3 | 2 | 2 | 1 |
| Primary school (%) | 85.71 | 88.16 | 2 | 3 | 2 | 2 | 1 |
| Secondary school (%) | 69.78 | 69.83 | 2 | 3 | 2 | 1 | 2 |
| University (%) | 71.49 | 73.62 | 1 | 3 | 2 | 2 | 2 |
Table 4. Weighting coefficients for language and visuospatial domains by educational group.
| Group | Naming | Repetition | Execution ability | Reading | Writing | Spatial structure |
|---|---|---|---|---|---|---|
| Illiterate (%) | 3 | 3 | 2 | 3 | 2 | 3 |
| Primary school (%) | 1 | 2 | 3 | 2 | 2 | 1 |
| Secondary school (%) | 2 | 2 | 2 | 1 | 1 | 1 |
| University (%) | 2 | 3 | 1 | 1 | 2 | 2 |
Table 5 summarizes the diagnostic performance before and after dynamic weighting, including accuracy, sensitivity, specificity, AUC, and the statistical significance of performance changes. The largest improvement occurred in the illiterate group (Δ=+7.25%), whereas the secondary school group showed the smallest improvement (Δ=+0.05%).
Table 5. Comparison of diagnostic performance metrics and statistical significance before and after dynamic weighting.
| Education group and metric | Before weighting | After weighting | Improvement (Δ) | P value |
|---|---|---|---|---|
| Illiterate | 7.25% | .06 | ||
| Accuracy (%) | 78.83 | 86.08 | ||
| Sensitivity (%) | 81.45 | 84.92 | ||
| Specificity (%) | 76.21 | 87.24 | ||
| AUCa | 0.835 | 0.912 | ||
| Primary school | 2.45% | .03 | ||
| Accuracy (%) | 85.71 | 88.16 | ||
| Sensitivity (%) | 86.10 | 88.54 | ||
| Specificity (%) | 85.32 | 87.78 | ||
| AUC | 0.894 | 0.928 | ||
| Secondary school | 0.05% | .46 | ||
| Accuracy (%) | 69.78 | 69.83 | ||
| Sensitivity (%) | 70.25 | 70.31 | ||
| Specificity (%) | 69.31 | 69.35 | ||
| AUC | 0.752 | 0.753 | ||
| University | 2.13% | .18 | ||
| Accuracy (%) | 71.49 | 73.62 | ||
| Sensitivity (%) | 72.15 | 74.88 | ||
| Specificity (%) | 70.83 | 72.36 | ||
| AUC | 0.768 | 0.795 |
AUC: area under the curve.
Weight distributions differed across education levels. For basic cognitive domains, spatial orientation weights were uniformly set to 3 across all groups, and immediate memory weights remained at 2. In the illiterate group, naming, repetition, and spatial construction received the highest weights (all=3). In the primary school group, temporal orientation was weighted at 2 and spatial construction was weighted at 1. In the secondary school group, executive function and calculation were weighted at 2, while spatial construction remained at 1. In the university group, executive function received the highest weight (3), calculation was weighted at 2, and spatial construction, reading, and writing each received a weight of 1.
Discussion
Principal Findings
This study developed an education-sensitive optimization framework for the MMSE using a dynamic SVM-based weighting strategy. Across 812 education-stratified participants, the model demonstrated that cognitive subitems contribute differently to diagnostic prediction depending on years of formal education. Dynamic weighting enhanced the performance of unimodal MMSE classifiers, particularly in low-literacy groups, and reduced education-related diagnostic bias. These findings refine the diagnostic utility of condensed cognitive screening tools and support their broader applicability in resource-limited settings.
Comparison With Prior Work
ML has been widely applied to cognitive assessment [32-35], yet existing studies typically use MMSE scores as features or labels without addressing education-driven measurement bias [36,37]. Prior condensed MMSE-only classifiers generally achieve accuracies of 72% to 85%, whereas multimodal MRI-, PET-, and biomarker-based models often exceed 90% [32,33]. Although our model remains unimodal, the dynamic weighting mechanism improved prediction by amplifying high-value signals and attenuating education-dependent interference, partially narrowing the performance gap with multimodal systems while sustaining low cost and operational simplicity.
Traditional MMSE optimization approaches—such as linear education corrections or standard ML models—lack the capacity to capture nonlinear, education-specific feature interactions [38]. In contrast, our framework integrates item deletion experiments with SVM-informed dynamic weighting to generate transparent, subgroup-adaptive coefficients. This approach offers an interpretable alternative to black box ensemble models and provides an effective mechanism for addressing item-level educational bias in cognitive assessment. A comprehensive comparison between the proposed dynamic weighted SVM model and existing cognitive assessment optimization strategies is presented in Table 6.
Table 6. Comparison of the proposed dynamic weighted Support Vector Machine (SVM) model with existing cognitive assessment optimization strategies.
| Optimization strategy | Key methodology | Data requirements | Advantages | Limitations |
|---|---|---|---|---|
| Traditional linear correction (eg, MMSEa-E) | Linear regression and fixed point addition | MMSE scores+demographics | Simple calculation and clinically familiar | Ignored nonlinear “cognitive interference” and low precision |
| Standard machine learning (eg, RFb, ANNc) | Black box classification using raw scores | MMSE subitems or total scores | High classification accuracy | Low interpretability (“black box”) and hard to explain to clinicians |
| Multimodal fusion models | Deep learning (CNNd/RNNe) integration | MRIf/PETg imaging+biomarkers+scale | Highest accuracy (>90%) and comprehensive pathology mapping | High cost, low accessibility in primary care, and complex deployment |
| This study (dynamic weighted SVM) | Nonlinear dynamic weighting via SVM | MMSE subitems only (low cost) | High interpretability (visible weights), education adaptive, and high accessibility | Accuracy is lower than multimodal models (feature limitation) |
MMSE: Mini-Mental State Examination.
RF: random forest.
ANN: artificial neural network.
CNN: convolutional neural network.
RNN: recurrent neural network.
MRI: magnetic resonance imaging.
PET: positron emission tomography.
Interpretation of Education-Specific Patterns
Education-stratified analyses revealed distinct, cognitive reliance patterns. The illiterate group depended heavily on orientation and immediate memory, consistent with basic functional domains commonly preserved in low-literacy populations. The primary school group demonstrated changes in weighting for temporal orientation, whereas secondary school participants exhibited varied interference across multiple domains and minimal model improvement (Δ=0.05), suggesting high within-group heterogeneity. University-educated participants showed stronger reliance on executive functioning and calculation. These patterns align with prior evidence that cognitive performance becomes increasingly distributed and strategy dependent with higher educational attainment [39,40].
The dynamic weighting strategy provided incremental accuracy improvements across all educational cohorts while reducing false-positive rates in lower-education groups. Subitem deletion further allowed visualization of feature importance (Figure 2), conceptually analogous to Shapley Additive Explanations–based explainability, thereby enhancing the transparency of SVM decision processes. Clinical validation by senior rehabilitation physicians confirmed that weight distributions reflected plausible neurocognitive patterns.
External validation using an independent community dataset (n=314) demonstrated stable performance (overall accuracy: 82.48% and AUC: 0.88), with the highest accuracy observed in the illiterate group (88.89%). These results support the generalizability and robustness of the proposed dynamic weighting model and mitigate concerns about overfitting.
Limitations
This study has several limitations. First, the sample was derived from a single geographic region, leading to educational imbalance across subgroups and potentially limiting generalizability. Broader multicenter sampling is needed. Second, demographic confounders beyond age and gender—such as socioeconomic status, occupational complexity, and urban-rural residence—were not available in the dataset, restricting the ability to fully disentangle education effects from related social determinants. Third, although the MMSE provides valuable screening utility, ceiling effects in highly educated individuals limit sensitivity to early cognitive decline. Integrating additional modalities, such as the Montreal Cognitive Assessment or imaging markers, may complement the proposed model. Finally, the SVM model relied on grid-search hyperparameter tuning; future research may explore AutoML-based optimization to improve efficiency and predictive performance.
Conclusions
This study proposes a dynamic SVM-based weighting framework that enhances the diagnostic fairness of MMSE-based cognitive screening across diverse educational backgrounds. By quantifying item-level contributions and adapting subitem weights for each education group, the method addresses a longstanding source of measurement bias in cognitive assessment. The approach retains the accessibility and scalability of condensed cognitive screening tools while improving prediction accuracy and interpretability. These findings provide a practical foundation for developing equitable cognitive assessment strategies, particularly in resource-limited regions.
Acknowledgments
The authors sincerely acknowledge all individuals and institutions that contributed to this work. This research was jointly supported by the Health Research Program of Anhui Province (project number AHWJ2023A10123) and the Anhui Provincial Teaching Research Project (project number 2022jyxm1851) titled "Research and Implementation of a Standardized Training Curriculum System for Rehabilitation Therapists Based on Competency-Based Education." Special thanks are extended to the funding bodies for their support.
Qi Zhang is co-corresponding author (email: ustczhangqi@163.com; phone +86 17333070277).
Abbreviations
- AUC
area under the curve
- ML
machine learning
- MMSE
Mini-Mental State Examination
- MRI
magnetic resonance imaging
- PET
positron emission tomography
- SVM
support vector machine
Footnotes
Funding: This work was supported by the Anhui Provincial Teaching Research Project (grant 2022jyxm1851), the Health Research Program of Anhui (grant AHWJ2023A10123), and the Anhui Provincial Graduate Academic Innovation Project “Optimization and Prediction of Alzheimer’s Disease-Related Cognitive Scales Based on Machine Learning” (grant 2024xscx009). This research received no commercial sponsorship or financial support that could compromise its objectivity. All authors declare no financial or proprietary interests relevant to this study. The funding agencies provided only financial support and had no involvement in study design, data collection and analysis, manuscript preparation, or submission for publication. The research team assumes full responsibility for the study's independence and scientific integrity.
Authors’ Contributions: QL: Conceptualization, writing – original draft, formal analysis, data curation, methodology
CM: Conceptualization, writing – original draft, formal analysis, data curation, methodology
ML: Investigation, resources
SC: Data curation (clinical)
MY: Data curation (clinical and community)
LX: Data curation (clinical and community)
QZ: Supervision, project administration
MW: Supervision, project administration
All authors have reviewed and approved the final manuscript, agreeing to assume public responsibility for its content.
QZ is the co-corresponding author.
Data Availability: The clinical dataset used in this study was obtained under ethical approval (2024-RE-431) and contains sensitive patient information; therefore, it cannot be made publicly available. Deidentified clinical records (n=404) may be shared upon reasonable request to the corresponding author under a data sharing agreement. Community screening data (n=408) are available with permission from the participating community health centers.
The complete source code for data preprocessing, Support Vector Machine modeling, dynamic weighting, and statistical analysis is openly available on GitHub [41].
Conflicts of Interest: None declared.
References
- 1.Chun CT, Seward K, Patterson A, Melton A, MacDonald-Wicks L. Evaluation of available cognitive tools used to measure mild cognitive decline: a scoping review. Nutrients. 2021 Nov 8;13(11):3974. doi: 10.3390/nu13113974. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Folstein MF, Folstein SE, McHugh PR. “Mini-mental state”. A practical method for grading the cognitive state of patients for the clinician. J Psychiatr Res. 1975 Nov;12(3):189–198. doi: 10.1016/0022-3956(75)90026-6. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 3.Jia X, Wang Z, Huang F, et al. A comparison of the Mini‑Mental State Examination (MMSE) with the Montreal Cognitive Assessment (MoCA) for mild cognitive impairment screening in Chinese middle‑aged and older population: a cross‑sectional study. BMC Psychiatry. 2021 Oct 4;21(1):485. doi: 10.1186/s12888-021-03495-6. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Patnode CD, Perdue LA, Rossom RC, et al. Screening for cognitive impairment in older adults: updated evidence report and systematic review for the US Preventive Services Task Force. JAMA. 2020 Feb 25;323(8):764–785. doi: 10.1001/jama.2019.22258. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 5.Scheltens P, De Strooper B, Kivipelto M, et al. Alzheimer’s disease. Lancet. 2021 Apr 24;397(10284):1577–1590. doi: 10.1016/S0140-6736(20)32205-4. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Yin TT, Cao MH, Yu JC, et al. T1‑weighted imaging‑based hippocampal radiomics in the diagnosis of Alzheimer’s disease. Acad Radiol. 2024 Dec;31(12):5183–5192. doi: 10.1016/j.acra.2024.06.012. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 7.Cardoso S, Barros R, Marôco J, de Mendonça A, Guerreiro M. Different MMSE domains are associated to cognitive decline and education. Appl Neuropsychol Adult. 2024;31(4):533–539. doi: 10.1080/23279095.2022.2041018. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 8.Bird HR, Canino G, Stipec MR, Shrout P. Use of the Mini‑Mental State Examination in a probability sample of a Hispanic population. J Nerv Ment Dis. 1987 Dec;175(12):731–737. doi: 10.1097/00005053-198712000-00005. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 9.Koepsell TD, Kurland BF, Harel O, Johnson EA, Zhou XH, Kukull WA. Education, cognitive function, and severity of neuropathology in Alzheimer disease. Neurology. 2008 May 6;70(19 Pt 2):1732–1739. doi: 10.1212/01.wnl.0000284603.85621.aa. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 10.Baker DW, Gazmararian JA, Sudano J, Patterson M. The association between age and health literacy among elderly persons. J Gerontol B Psychol Sci Soc Sci. 2000 Nov;55(6):S368–S374. doi: 10.1093/geronb/55.6.s368. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 11.Ouvrard C, Berr C, Meillon C, et al. Norms for standard neuropsychological tests from the French CONSTANCES cohort. Eur J Neurol. 2019 May;26(5):786–793. doi: 10.1111/ene.13890. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 12.Aprahamian I, Martinelli JE, Cecato J, Yassuda MS. Screening for Alzheimer’s disease among illiterate elderly: accuracy analysis for multiple instruments. J Alzheimers Dis. 2011;26(2):221–229. doi: 10.3233/JAD-2011-110125. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 13.Rezende GP, Cecato J, Martinelli JE. Cognitive abilities screening instrument-short form, Mini-Mental State Examination and functional activities questionnaire in the illiterate elderly. Dement Neuropsychol. 2013;7(4):410–415. doi: 10.1590/S1980-57642013DN74000009. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Katzman R, Zhang MY, et al. A Chinese version of the Mini-Mental State Examination; impact of illiteracy in a Shanghai dementia survey. J Clin Epidemiol. 1988;41(10):971–978. doi: 10.1016/0895-4356(88)90034-0. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 15.Weiss BD, Reed R, Kligman EW, Abyad A. Literacy and performance on the Mini-Mental State Examination. J Am Geriatr Soc. 1995 Jul;43(7):807–810. doi: 10.1111/j.1532-5415.1995.tb07057.x. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 16.Paddick SM, Gray WK, McGuire J, Richardson J, Dotchin C, Walker RW. Cognitive screening tools for identification of dementia in illiterate and low-educated older adults, a systematic review and meta-analysis. Int Psychogeriatr. 2017 Jun;29(6):897–929. doi: 10.1017/S1041610216001976. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 17.Pellicer-Espinosa I, Díaz-Orueta U. Cognitive screening instruments for older adults with low educational and literacy levels: a systematic review. J Appl Gerontol. 2022 Apr;41(4):1222–1231. doi: 10.1177/07334648211056230. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Brucki SMD, Mansur LL, Carthery-Goulart MT, Nitrini R. Formal education, health literacy and Mini-Mental State Examination. Dement Neuropsychol. 2011;5(1):26–30. doi: 10.1590/S1980-57642011DN05010005. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Carnero-Pardo C. Should the Mini-Mental State Examination be retired? Neurologia. 2014 Oct;29(8):473–481. doi: 10.1016/j.nrl.2013.07.003. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 20.Cai H, Shao Y, Liu XY, et al. Interpretable prognostic modeling for long-term survival of Type A aortic dissection patients using Support Vector Machine algorithm. Eur J Med Res. 2025 Apr 15;30(1):277. doi: 10.1186/s40001-025-02510-w. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Luo W. Predicting cervical cancer outcomes: statistics, images, and machine learning. Front Artif Intell. 2021;4:627369. doi: 10.3389/frai.2021.627369. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Cai F, Cherkassky V. Generalized SMO algorithm for SVM-based multitask learning. IEEE Trans Neural Netw Learn Syst. 2012 Jun;23(6):997–1003. doi: 10.1109/TNNLS.2012.2187307. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 23.Huang C, Chung FL, Wang S. Multi-view L2-SVM and its multi-view core vector machine. Neural Netw. 2016 Mar;75:110–125. doi: 10.1016/j.neunet.2015.12.004. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 24.Mastropietro A, Feldmann C, Bajorath J. Calculation of exact Shapley values for explaining Support Vector Machine models using the radial basis function kernel. Sci Rep. 2023 Nov 10;13(1):19561. doi: 10.1038/s41598-023-46930-2. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Jasodanand VH, Bellitti M, Kolachalama VB. An AI-first framework for multimodal data in Alzheimer’s disease and related dementias. Alzheimers Dement. 2025 Sep;21(9):e70719. doi: 10.1002/alz.70719. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Rathore S, Habes M, Iftikhar MA, Shacklett A, Davatzikos C. A review on neuroimaging-based classification studies and associated feature extraction methods for Alzheimer’s disease and its prodromal stages. Neuroimage. 2017 Jul 15;155:530–548. doi: 10.1016/j.neuroimage.2017.03.057. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Zhang D, Wang Y, Zhou L, Yuan H, Shen D. Multimodal classification of Alzheimer’s disease and mild cognitive impairment. Neuroimage. 2011 Apr 1;55(3):856–867. doi: 10.1016/j.neuroimage.2011.01.008. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Jasodanand VH, Kowshik SS, Puducheri S, et al. AI-driven fusion of multimodal data for Alzheimer’s disease biomarker assessment. Nat Commun. 2025 Aug 11;16(1):7407. doi: 10.1038/s41467-025-62590-4. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Dincer A, Gordon BA, Hari-Raj A, et al. Comparing cortical signatures of atrophy between late-onset and autosomal dominant Alzheimer disease. Neuroimage Clin. 2020;28:102491. doi: 10.1016/j.nicl.2020.102491. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Qiu S, Miller MI, Joshi PS, et al. Multimodal deep learning for Alzheimer’s disease dementia assessment. Nat Commun. 2022 Jun 20;13(1):3404. doi: 10.1038/s41467-022-31037-5. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Castellano G, Esposito A, Lella E, Montanaro G, Vessio G. Automated detection of Alzheimer’s disease: a multi-modal approach with 3D MRI and amyloid PET. Sci Rep. 2024 Mar 3;14(1):5210. doi: 10.1038/s41598-024-56001-9. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Vyas A, Aisopos F, Vidal ME, Garrard P, Paliouras G. Identifying the presence and severity of dementia by applying interpretable machine learning techniques on structured clinical records. BMC Med Inform Decis Mak. 2022 Oct 17;22(1):271. doi: 10.1186/s12911-022-02004-3. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Feng F, Wang P, Zhao K, et al. Radiomic features of hippocampal subregions in Alzheimer’s disease and amnestic mild cognitive impairment. Front Aging Neurosci. 2018;10:290. doi: 10.3389/fnagi.2018.00290. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Battista P, Salvatore C, Castiglioni I. Optimizing neuropsychological assessments for cognitive, behavioral, and functional impairment classification: a machine learning study. Behav Neurol. 2017;2017:1850909. doi: 10.1155/2017/1850909. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Khatun S, Morshed BI, Bidelman GM. A single‑channel EEG‑based approach to detect mild cognitive impairment via speech‑evoked brain responses. IEEE Trans Neural Syst Rehabil Eng. 2019 May;27(5):1063–1070. doi: 10.1109/TNSRE.2019.2911970. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Wu Y, Jia M, Xiang C, Lin S, Jiang Z, Fang Y. Predicting the long-term cognitive trajectories using machine learning approaches: a Chinese nationwide longitudinal database. Psychiatry Res. 2022 Apr;310:114434. doi: 10.1016/j.psychres.2022.114434. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 37.Martin SA, Townend FJ, Barkhof F, Cole JH. Interpretable machine learning for dementia: a systematic review. Alzheimers Dement. 2023 May;19(5):2135–2149. doi: 10.1002/alz.12948. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Kantayeva G, Lima J, Pereira AI. Application of machine learning in dementia diagnosis: a systematic literature review. Heliyon. 2023 Nov;9(11):e21626. doi: 10.1016/j.heliyon.2023.e21626. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Livingston G, Huntley J, Sommerlad A, et al. Dementia prevention, intervention, and care: 2020 report of the Lancet Commission. Lancet. 2020 Aug 8;396(10248):413–446. doi: 10.1016/S0140-6736(20)30367-6. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Stern Y. Cognitive reserve in ageing and Alzheimer’s disease. Lancet Neurol. 2012 Nov;11(11):1006–1012. doi: 10.1016/S1474-4422(12)70191-6. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Dynamic-SVM-weighting-for-MMSE-optimization. GitHub. [05-01-2026]. https://github.com/masC1962/Dynamic-SVM-Weighting-for-MMSE-Optimization URL. Accessed.

