Skip to main content
BMC Musculoskeletal Disorders logoLink to BMC Musculoskeletal Disorders
. 2026 Jan 8;27:52. doi: 10.1186/s12891-025-09385-2

Risk prediction models for postmenopausal osteoporosis: a systematic review and meta-analysis study

Lingjia Li 1,2, Xiangzhou Lan 1, Weike Zeng 3, Yujiao Xu 4, Qing Chen 1,
PMCID: PMC12829196  PMID: 41507834

Abstract

Background

Postmenopausal osteoporosis usually happens 5 ~ 10 years after menopause. Low awareness, low detection rates and high morbidity have prevented the possibility of early or preventive interventions, thus increasing the social and economic burden on families and societies. A reliable prediction model for postmenopausal osteoporosis has the potential to guide the prevention, but regarding the early prediction of postmenopausal osteoporosis without fracture, this field has not been sufficiently studied. Although many scholars have developed several prediction models to estimate the risk of postmenopausal osteoporosis without fractures, the evidence about the model quality and clinical applicability is scarce.

Method

Nine databases (Medline, Embase, Web of science, CINAHL, The Cochrane Library, CNKI, SinoMed, Wanfang, VIP data) were systematically searched from 1 January 2014 to 1 May 2024. Two researchers independently extracted data using the CHARMS checklist and assessed bias using the PROBAST tool. The primary outcomes of interest were related to the model’s discriminative ability (assessed by pooled AUC values) and calibration performance (evaluated using calibration curves or the calibration intercept and slope). We performed meta-regression and sensitivity analyses to explore the influence of important factors, such as data sources, machine learning methods, and types of predictor variables, on the aforementioned. results. Additionally, subgroup analyses were conducted based on data sources, machine learning. methods, and types of predictor variables. The study was registered in the PROSPERO database (registration number CRD42024542498).

Results

A total of 8,549 records were initially identified, and 7 studies (comprising 19 models) were ultimately included. All models were developed based on Asian population data. The risk of bias assessment showed: 1 study had a low risk, 1 study had an unclear risk, and 5 studies had a high risk. The sample sizes ranged from 319 to 4,417 participants. The reported AUC of the models ranged from 0.639 to 0.921; however, the vast majority of studies lacked reports on calibration performance. The pooled C-statistic (AUC) was 0.78 (95%CI: 0.73–0.83). Sensitivity analysis yielded robust results (AUC=0.77). Subgroup analysis indicated that models combining demographic and laboratory data demonstrated the best performance (AUC=0.92). Significant publication bias and substantial heterogeneity (I² = 98%) were observed among the studies.

Conclusion

Current machine learning-based prediction models for postmenopausal osteoporosis without fractures, as presented in the included studies, demonstrate good discriminative ability but are generally characterized by a high risk of bias, a notable lack of calibration performance evaluation, and insufficient validation of clinical utility. Furthermore, existing models are developed entirely on Asian population data, which limits their generalizability to other populations. Future research should focus on strictly adhering to prediction model research guidelines (such as PROBAST), enhancing the reporting of model calibration and clinical utility, and assessing model generalizability through external validation in multi-center studies encompassing diverse ethnicities and regions.

Supplementary Information

The online version contains supplementary material available at 10.1186/s12891-025-09385-2.

Keywords: Machine learning, Postmenopausal osteoporosis, Systematic review

Introduction

Osteoporosis (OP) is a metabolic bone disease characterized by reduced bone density, degeneration of the trabecular structure and increased risk of fractures. Osteoporosis-related fractures impose a substantial socioeconomic burden worldwide. Annual expenditures reach approximately $1.79 billion in the United States and £4 billion in the United Kingdom [1]. Globally, the prevalence of osteoporosis remains alarmingly high among women, with an estimated rate of 23.1%. Postmenopausal women are particularly vulnerable due to a dramatic decline in estrogen levels caused by ovarian function deterioration, which significantly disrupts bone metabolism and places them in the highest-risk category [2]. Postmenopausal osteoporosis (PMOP) affects approximately 200 million women worldwide, with significant regional variations in prevalence. Morocco reports a pooled prevalence rate of 32% [3]. In Asia, particularly in populous countries like China, the prevalence of OP among women aged 50 and above has been reported to be significantly higher than in some European and American countries [4]. However, a key global challenge remains the widespread lack of awareness about PMOP among patients. Regional surveys indicate that in some Asian countries, the awareness rate of OP among women aged 40–49 is as low as 0.9% [5], although more comprehensive and updated data are still needed. Early prevention and timely diagnosis are crucial for maintaining bone mineral density and slowing disease progression. According to the World Health Organization (WHO) diagnostic criteria for osteoporosis [6], the bone density measured compared to the peak bone density of the same sex, standard deviation of the bone density decreased (BMD) by T value ≤ −2.5 can be diagnosed as osteoporosis. However, the accessibility and utilization of BMD testing are generally inadequate. Studies have shown that the rate of BMD testing among postmenopausal women after sustaining a distal radius fracture remains low and has shown a declining trend, once being as low as 10.5% [7]. This issue is even more severe in the general population and in resource-limited settings (e.g., reports indicate that the testing rate in China was once as low as 3.7%) [5]. Low awareness, low detection and high prevalence have resulted in most patients not taking timely preventive and control measures in the early stages of bone loss until pain, spinal deformation and fractures occurring, which prevents the possibility of early or preventive interventions, thus increasing the social and economic burden on families and societies. A range of prescreening tools are developed to predict the risk of osteoporosis, such as the osteoporosis self-assessment tool for Asians [8], the osteoporosis risk assessment instrument [9], simple calculated osteoporosis risk estimation [10] and so on. However, these tools exhibit significant limitations in predictive performance due to their oversimplification of the complex pathophysiology of bone metabolism. Their area under the receiver operating characteristic curve (AUROC) typically ranges between 0.65 and 0.75 [1113], which substantially restricts their application in high-precision risk stratification and poor predictive performance leads to missed opportunities for early intervention. Moreover, they demonstrate limited capacity to effectively incorporate the increasingly available clinical data, particularly serum biochemical markers that are closely associated with bone metabolism.

In recent years, with the accumulation of medical big data and advancements in artificial intelligence, machine learning (ML) technology has demonstrated strong potential in medical prediction fields. The core advantage of ML models lies in their ability to learn nonlinear relationships and complex interactions between features, potentially overcoming the limitations of traditional tools. By integrating multidimensional data including demographic information, clinical variables, and key serum bone metabolism markers (such as bone turnover markers, vitamin D, and hormone levels), machine learning models provide a new opportunity to develop more accurate risk prediction tools for the PMOP population without fractures. However, current research still exhibits significant gaps in external validation and clinical generalizability of these models. Tartibian B et al. [14] employed the k-nearest neighbors (KNN) algorithm to predict osteoporosis risk in elderly women, achieving 61.7% accuracy in identifying femoral neck OP risk. The study proposed this method could serve as a preliminary screening tool prior to DXA scans, thereby reducing unnecessary testing. Subsequently, Fasihi L et al. [15] systematically compared eight algorithms and demonstrated that gradient boosting (GB) and random forest (RF) outperformed KNN for OP prediction, while Ada Boost proved most suitable for generating exercise prescriptions. Notably, the study developed an innovative “prediction-intervention” closed-loop system that directly outputs exercise protocols (e.g., recommending aquatic training combined with balance exercises for osteoporotic women), effectively shortening clinical decision-making pathways and addressing personalized healthcare needs. It must be pointed out that the safety and appropriateness of personalized exercise programs generated by such models still require evaluation and confirmation by trained exercise or medical professionals. These studies demonstrate that ML models exhibit good discriminative ability for OP risk prediction and possess potential as clinical decision-support tools.

Simultaneously, the development of OP risk prediction models faces significant challenges. Wu et al. [16] evaluated 53 osteoporosis prediction models involving 15,209,268 patients. Their study revealed that while machine learning demonstrated good performance in fracture prediction (with relatively high pooled C-indices), the vast majority of models lacked adequate external validation to confirm their reliability and generalizability. This insufficient validation landscape substantially restricts the reliable application of models across diverse populations and clinical settings, constituting a critical bottleneck hindering the clinical translation of ML technologies. The development of rigorously externally validated PMOP-specific prediction models with robust generalizability has thus emerged as a pivotal scientific challenge in the field. Compared to the study by Wu Yu et al., which primarily evaluated fracture risk prediction in diagnosed patients, the core innovation of this research lies in its precise focus on the early risk prediction of PMOP without fracture, which is a critical yet under explored area of study.

Therefore, this study aims to systematically evaluate the current research status of machine learning models in predicting the risk of PMOP without fracture. The study will emphasize analyzing the predictive performance of existing models, rigorously assessing their external validation status and evidence of clinical generalizability, and systematically introducing a comprehensive review of model reporting quality and methodological bias risks. Through this systematic evaluation, the study aims to provide evidence-based insights and optimization strategies for developing more robust and clinically translatable risk prediction tools of PMOP without fracture, thereby facilitating early identification and intervention in high-risk populations and ultimately reducing the incidence and disease burden of this major public health issue.

Method

This systematic review and meta-analysis were according to the preferred reporting items suggested by the Cochrane Collaboration. The meta-analysis was registered in the PROSPERO database (registration number CRD42024542498). The study protocol explicitly defined the research questions, search strategy, inclusion/exclusion criteria, data extraction items, risk of bias assessment methods, and planned analytical approaches. The final version was publicly available in the PROSPERO registry record or could be obtained by contacting the corresponding author.

Search strategy

We conducted a systematic search of the literature from Medline, Embase, Web of science, CINAHL, The Cochrane Library, CNKI, SinoMed, Wanfang, VIP data. Search time ranged from 1 January 2014 to 1 May 2024. Articles were searched in databases by combining related terms such as “postmenopausal osteoporosis”, “bone loss”, “machine learning”, “prediction model”. We applied a prediction model-specific filter (Hayden filter [17]) and manually traced reference lists. The complete search strategy is available in Supplementary File 1 (S1 search strategy and S2 study selection criteria). As this study was based on previously published research, it was not applicable for permissible consent from participants.

Inclusion and exclusion criteria

Inclusion criteria as follows

This study established the inclusion criteria based on the PICOS principle, as follows: (1)Population (P): Postmenopausal women. (2)Study Design (I): Cohort studies or case-control studies that involved the development and/or validation of prediction models for postmenopausal osteoporosis without fractures. (3)Prediction Model (C): The core of the study must be the development, validation, or updating of a prediction model. (4)Outcome (O): The primary outcome is the diagnosis of postmenopausal osteoporosis (based on a bone density T-score ≤ −2.5 as measured by DXA. (5) Model Performance (S): The study must report at least one model performance metric.

Exclusion criteria as follows

(1) the content of the study only referred to predictors or risk factors, no predictive model had been established; (2) used qualitative methods to construct prediction models; (3) reported repeatedly, abstracts, conference papers, cases, and reviews; (4) Non-Chinese and English.

Data extraction

Two researchers (L.J.L. and X.Z.L.) independently extracted the data from the included studies. For data discrepancies, two independent reviewers will first resolve disagreements through discussion. If consensus cannot be reached, a third senior investigator (Y.J.X.) will arbitrate the final decision. For studies with missing original data, we tried to contact the author to request the data. If no response was received, we would attempt to derive the required data through reported results (e.g., calculating standard errors from confidence intervals [18]). Studies would only be excluded when key performance metrics were completely unavailable and cannot be reasonably estimated, with all exclusion reasons being systematically documented. For studies reporting multiple models within the same publication, we extracted performance data from all eligible models reported in that study. For instances where multiple versions of the same model were reported across different publications, we prioritized the inclusion of the final version, the one with the most complete set of predictors and the most thorough validation to avoid duplicate counting of the same model population. The information extracted was according to the checklist for critical appraisal and data extraction for systematic reviews of prediction modeling studies (CHARMS) [19]. The details were as follows: (1) Basic study information: first author, publication year, country, data source. (2) Study population and outcome: sample size, number of osteoporosis events, sample size adequacy (the events per variable, EPV) [20]. (3) Predictors: list of candidate predictors, predictors finally included in the model, type of predictors (demographic characteristics primarily including age, BMI, age at menopause, height, etc., laboratory indicators primarily biochemical and serum markers, other indicators primarily lifestyle factors, medication use, and nutritional intake). (4) Model development methods: predictor selection method, method for handling missing data, model algorithm (e.g., Logistic Regression, RF, etc.), type of validation. (5) Model performance metrics: ①Discrimination: Refers to the model’s ability to distinguish between individuals with and without the disease. Extracted metrics include the C-statistic/area under the receiver operating characteristic curve (AUC), sensitivity, specificity, etc. ② Calibration: Refers to the agreement between the predicted risk by the model and the actual observed risk. Extracted metrics include p-value of the Hosmer-Lemeshow test, calibration slope, etc. ③ Other metrics: accuracy, positive predictive value (PPV), negative predictive value (NPV), and Youden’s index were also extracted. ④ Clinical utility: Record whether decision curve analysis (DCA) was reported to evaluate the model’s net clinical benefit. (6) Model presentation: Extract the final presentation format of the model (e.g., nomogram, online calculator, etc.).

Risk of bias assessment

Prediction model risk of bias assessment tool (PROBAST) [21] was used to assess the bias risk of models. Two reviewers (L.J.L. and X.Z.L.) used the tool to assess each article independently from four domains: study subjects, predictors, outcomes and statistical analysis. During assessment, each question was answered as yes, probably yes, probably no, no, or no information, with yes indicating a low risk of bias and no indicating a high risk of bias. After completing independent assessments, the reviewers would enter their results into standardized forms for comparison. For any discrepancies identified, the two reviewers would first discuss and attempt to reach consensus. If disagreements persist after discussion, a third senior investigator (Y.J.X.) would review the relevant evidence and make the final determination.

Data analysis

We used a random effects model to pool AUCs across studies and presented the results in forest plots. The degree of inter-study heterogeneity using the Cochran’s Q test [22] (where significant heterogeneity was defined as P ≤ 0.10 or > 50%) to assess whether a fixed effects model could have been used. We conducted meta-regression analyses to investigate potential sources of heterogeneity and performed subgroup analyses to examine the consistency of effects based on data sources (database vs. electronic medical record), machine learning methods (LR vs. others), and types of predictors (demographics vs. demographics and other indicators vs. demographics, laboratory and others), with a predefined significance threshold of P < 0.05. Sensitivity analysis was performed to further identify the source of heterogeneity by removing each study and re-calculating the pooled effect size of the remaining studies. Publication bias was assessed using Egger’s linear regression test, with a significance threshold of P < 0.05 indicating potential bias. In cases of funnel plot asymmetry, the trim-and-fill method was applied to adjust for potential missing studies. The meta analysis was performed using the software R V.4.2.2 (R Development Core Team, Vienna, http://www.R-project.org). It was considered statistically significant that p value less than 0.05.

Results

Study selection

A total of 8549 studies were identified from nine databases and other sources. After removing 1775 duplicates and 62 records marking as ineligible by manual reviewing, 6172 articles remained. Through screening the title and abstract of the article, the remaining 28 records were reviewed by full-text screening, among which 12 records had not established prediction model, 4 records were qualitative study, 2 records were unable to access the full data, 3 records included additional outcome indicators. Finally, 7 articles were included in this study (Fig. 1).

Fig. 1.

Fig. 1

Flow chart of study selection

Study characteristics

7 studies yielded 19 models and published from 2020 to 2023, with four [2426, 29] published in 2023. The 7 included models were from four different continents, including China (n = 4, 57.1%), Korea (n = 2, 28.6%), and Thailand (n = 1, 14.3%). Most of prediction models (n = 5, 71.4%) were developed using data from retrospective studies. Sample sizes included in the model were between 319 and 4417 cases, and patients with postmenopausal osteoporosis ranged from 157 to 837. Study characteristics were summarized in Table 1.

Table 1.

Study characteristics (n = 7)

Study Year Country Data source Number of participants/number
of events (EPV)
Shim JG et al [23] 2020 Korean Retrospective study 1792/613 (EPV<20)
Wu Y et al [24] 2023 China Retrospective study 4417/837 (EPV<20)
Wang SR et al [25] 2023 China Case-control 323/185 (EPV<20)
Liu GK et al [26] 2023 China Retrospective study 350/157 (EPV<20)
Kwon Y et al [27] 2022 Korean Retrospective study 1431/821 (EPV<20)
Makond B et al [28] 2022 Thailand Retrospective study 356/266 (EPV<20)
Wang JL et al [29] 2023 China Case-control 319/159 (EPV<20)

Model build and predictive performance

The number of predictors included in the final model predictors ranged from 3 to 20 and the main feature categories selected for building the predictive model were demographic characteristics (primarily including age, BMI, age at menopause, height, etc.), laboratory indicators (primarily biochemical and serum markers), other indicators (primarily lifestyle factors, medication use, and nutritional intake). The final inclusion of the predictors in the 7 models was detailed in Supplementary File 1 (Figure S3 Included predictors of each study). For predictor selection final to modeling, the most frequently applied method was logistic regression [23, 24, 26, 29]. As for the way to handle missing data, 3 studies [25, 26, 29] did not report missing values, other studies were imputed using multiple imputation or eliminated directly (Table 2).

Table 2.

Model build and predictive performance (n = 7)

Study Number of candidate predictors/predictors included in model Type of predictors Predictor selection Missing data
Shim JG et al [23] 19/9

logistic regression

(backward stepwise)

directly eliminate
Wu YQ et al [24] 17/9 multivariate logistic regression

eliminate and

multiple imputation

Wang SR et al [25] 26/6 LASSO not reported
Liu GK et al [26] 15/5 logistic regression not reported
Kwon Y et al [27] NR/20 recursive feature elimination eliminate
Makond B et al [28] NR/11 NR K-nearest neighbor imputation
Wang JL et al [29] 6/3 logistic regression not reported

Abbreviations: ①demographic characteristics, ②demographic characteristics and other indicators, ③demographic characteristics, laboratory indicators and others

LASSO Least Absolute Shrinkage and Selection Operator, NR Not Reported

As for model development, the logistic regression method was commonly used except for Makond B et al.’s [28] study, in which decision trees machine learning methods were used to generate prediction models, and Kwon Y et al. [27] applied other 3 machine learning methods (random forest, AdaBoost, gradient boosting machine). On the model validation, among the six studies [2328] that employed internal validation of which 2 studies [27, 28] reported an 8:2 training-test split, 1 study [23] used a 7:3 split ratio. The remaining 3 studies [2426] failed to specify their validation methodology. Only Wang JL et al.’s [29] study was externally validated using temporal validation. The model presentation format was reported in 4 models (57.1%), of which all models were presented as nomogram. Specific information provided in Table 3.

Table 3.

Model build and predictive performance (n = 7)

Study Model development methods Type of validation Model presentation
Shim JG et al [23] LR, KNN, DT, RF, GBM, SVM, ANN k-fold cross-validation not reported
Wu YQ et al [24] LR, Bayesian sample splitting Nomogram
Wang SR et al [25] LR internal validation Nomogram
Liu GK et al [26] LR sample splitting Nomogram
Kwon Y et al [27] RF, AdaBoost, GBM k-fold cross-validation not reported
Makond B et al [28] CART, QUEST, CHAID, C4.5 10-fold cross-validation not reported
Wang JL et al [29] LR external validation Nomogram

Abbreviations: LR Logistic Regressio, KNN K-Nearest Neighbors, DT Decision Tree, RF Random Forest, GBM Gradient Boosting Machine, SVM Support Vector Machines, ANN Artificial Neural Network, CART Classification and Regression Tree, QUEST Quick, Unbiased, Efficient Statistical Tree, CHAID Chi-squared Automatic Interaction Detection

In model validation, the reported C-index ranged from 0.639 to 0.921. Regarding model calibration performance, only three studies [24, 25, 29] provided both calibration curves and reported the results of the Hosmer-Lemeshow test (all p-values > 0.05), while one study [26] provided a calibration curve. For other performance metrics, four studies [23, 24, 26, 28] reported accuracy (ranging from 0.644 to 0.849), and four studies [23, 24, 28, 29] reported sensitivity (0.580 to 0.875) and specificity (0.560 to 0.880). Most studies [2426, 28, 29] also reported metrics such as Youden’s index, positive predictive value, and negative predictive value, with some studies including decision curves (details are provided in Table 4).

Table 4.

Model performance characteristics(n = 7)

Study ML methods Model performance
AUROC
(95%CI)
Accuracy
(95%CI)
Sensitivity
(95%CI)
Specificity
(95%CI)
Other indicators

Shim JG

et al [23]

LR 0.727(0.672–0.753) 0.749(0.706–0.789) 0.79(0.72–0.85) 0.66(0.60–0.72) -
KNN 0.713(0.687–0.778) 0.747(0.704–0.787) 0.58(0.50–0.66) 0.85(0.80–0.89)
DT 0.685(0.641–0.731) 0.706(0.661–0.748) 0.60(0.52–0.68) 0.77(0.71–0.82)
RF 0.734(0.688–0.773) 0.747(0.704–0.787) 0.68(0.61–0.75) 0.79(0.73–0.83)
GBM 0.728(0.672–0.755) 0.718(0.673–0.759) 0.77(0.70–0.83) 0.69(0.63–0.74)
SVM 0.728(0.674–0.758) 0.727(0.682–0.768) 0.73(0.66–0.80) 0.72(0.67–0.78)
ANN 0.743(0.693–0.777) 0.749(0.706–0.789) 0.72(0.64–0.79) 0.77(0.71–0.82)

Wu Y

et al [24]

LR 0.752(0.734–0.768) 0.644 0.762 0.616

H-L

(P = 0.125), calibration curve (slope value ≈ 1)

Jorden index

Bayesian 0.764(0.747–0.780) 0.678 0.708 0.672
Wang SR et al [25] LR 0.915(0.876–0.954) - - -

H-L (P > 0.05),

calibration curve

decision curve

Liu GK

et al [26]

LR

AUROC1=

0.792(0.745–0.840)

AUROC2=

0.814(0.756–0.870)

- - -

calibration curve

decision curve

Kwon Y

et al [27]

RF 0.919 0.832 - - -
AdaBoost 0.921 0.849
GBM 0.908 0.829
Makond B et al [28] CART 0.702 0.784 0.84 0.600

PPV = 0.815

NPV = 0.500

QUEST 0.773 0.722 0.667 0.880

PPV = 0.842

NPV = 0.375

CHAID 0.735 0.804 0.875 0.600

PPV = 0.828

NPV = 0.667

C4.5 0.639 0.711 0.764 0.560

PPV = 0.808

NPV = 0.444

Wang JL

et al [29]

LR 0.711(0.656–0.767) 0.742 0.581 -

calibration curve

decision curve

Abbreviations: AUROC1, training model, AUROC2, testing model, PPV Positive Predictive value, NPV Negative predictive value, HL Hosmer-Lemeshow goodness-of-fit test

Meta analysis

We merged the C statistics of each model (Fig. 2). The results showed that the pooled C-statistic for the postmenopausal osteoporosis prediction model was 0.78 (95% CI: 0.73, 0.83). There was a high degree of heterogeneity among the included studies (I² = 98.0%, τ² = 0.0071), with a 95% prediction interval ranging from 0.54 to 0.93, which primarily attributable to variations in data sources, modeling approaches, and type of predictors. Therefore, we conducted meta regression and subgroup analyses to explore the sources of heterogeneity.

Fig. 2.

Fig. 2

Forest plot of C statistics

Meta-regression analysis revealed that the type of predictors was a significant source of performance heterogeneity (see Supplementary File 1, Table S4 meta-regression analysis of subgroups). Subgroup analysis results indicated that models incorporating demographic, laboratory, and other indicators reported the highest AUC value (0.92), and this difference was statistically significant. It was important to emphasize that this subgroup with the high AUC value included only 2 studies with a small sample size. Data source and modeling method were not significant sources of heterogeneity (Supplementary File 1, Table S5, subgroup analysis).

In the leave-one-out sensitivity analysis, the meta-analysis results showed an AUC = 0.77 (95% CI:0.73–0.81, P < 0.001) with heterogeneity I²=98.0%. Compared to the overall meta- analysis result of all models (AUC = 0.78, 95%CI: 0.73–0.83), the results demonstrated that the pooled AUC remained relatively stable upon individual study removal in sensitivity analyses. The random-effects meta-analysis results were provided in Supplementary File 1 (Figure S6 sensitive analysis results).

This study assessed publication bias through Egger’s regression test (t = −3.67, P = 0.0019) and trim-and-fill analysis. The Egger’s test revealed significant small-study effects (bias estimate = −9.14, SE = 2.49), indicating funnel plot asymmetry with smaller studies clustered in higher AUC regions (Supplementary File 1 Figure S7 funnel plot and egger’s regression test). Following trim-and-fill adjustment for 9 theoretically missing studies, the pooled AUC decreased from the original estimate of 0.845 to 0.795. Notably, extreme heterogeneity persisted throughout (I² = 98%, tau² = 0.0174) without significant improvement post-adjustment (Supplementary File 1 Figure S8 trim-and-fill adjustment).

Risk of bias

A total of 7 studies were ultimately included, comprising 1 study with low risk of bias, 1 study with unclear risk of bias, and 5 studies with high risk of bias (Fig. 3). In the participants domain, the primary reason for high risk of bias was the retrospective study design (n = 5, 71.4%) [23, 24, 2628], which might have introduced bias in case selection and control group definition. In the predictors domain, most of studies with high risk of bias (n = 5, 71.4%) [23, 24, 2628] failed to report blinded assessment or standardized measurement methods for predictors and outcome indicators, potentially leading to measurement bias and expectation bias. In the analysis domain, the most common shortcomings were insufficient sample size (n = 7, 100%, EPV < 20) and risk of overfitting (n = 7, 100%). One additional study (n = 1, 14.3%) [26] was rated as high risk due to variable selection based solely on univariate analysis. This screening method could lead to the omission of important predictors and reduced model stability. The overall applicability of the 7 studies was better because the participants included were easy to find in the clinic and the relevant data of clinical diagnosis and treatment were also taken into account when selecting the variables.

Fig. 3.

Fig. 3

Risk of bias assessment. Abbreviations: Red:high risk; Yellow:unclear; Green:low risk [2329]

Discussion

This study conducted a literature search on risk prediction models for postmenopausal osteoporosis both domestically and internationally, ultimately including 7 studies with a total of 19 risk prediction models. All studies reported model discrimination (0.639–0.921), indicating that most models had moderate to excellent discrimination. However, three studies [23, 24, 27] failed to report calibration metrics, significantly limiting the evaluation of their clinical credibility. Calibration performance, which reflects the agreement between a model’s predicted probabilities and actual observed probabilities, serves as a critical indicator of a risk prediction model’s clinical utility [30]. The lack of calibration assessment may lead to several important issues. It prevents confirmation of whether the model’s predicted absolute risks are accurate - even with good discrimination, predicted probabilities may systematically overestimate or underestimate actual risks. It makes clinicians hesitant to trust the specific risk values provided by the model, thereby limiting its application in individualized decision-making [31, 32].

Furthermore, clinical utility represents the final critical step in evaluating the value of a prediction model. Among the 19 models included in this study, only three studies (encompassing 3 models in total) performed DCA to assess the models’ net benefit. DCA provides crucial evidence for practical application by quantifying the clinical benefit of using the model compared to default strategies (such as intervening on all or no patients) across various threshold probabilities [33]. The absence of DCA in the vast majority of studies means it is impossible to determine at which risk thresholds a model provides superior clinical value over simple strategies. This lack of evidence supporting the model’s ability to improve patient outcomes or enhance medical decision-making efficiency suggests that its so-called “high discriminative ability” may not translate into meaningful clinical net benefit. Similar to the lack of calibration assessment, the absence of clinical utility evaluation severely limits these models’ potential to evolve from mere statistical tools into reliable clinical decision support systems. Based on the above findings, we recommend that future developers of postmenopausal osteoporosis prediction models should be required to mandatorily report the model’s calibration performance (e.g., calibration plot, calibration slope) and clinical utility (e.g., DCA) to comprehensively demonstrate its clinical applicability. Furthermore, it is strongly recommended to strictly adhere to international guidelines such as PROBAST during the model development phase. This will help control the risk of bias across various aspects, including study design, data quality, analytical methods, and result reporting, thereby enhancing the model’s scientific rigor and generalizability potential.

Although external validation is crucial for assessing a model’s generalizability, this review found that only one study [27] performed external validation. This finding aligns with the results of Ramspek CL et al. [34], which indicated that only approximately 5–7% of prediction model studies complete external validation. This widespread lack of external validation creates significant uncertainty regarding the performance of numerous models in real-world clinical practice. It is particularly noteworthy that all prediction models included in this review originated from Asian countries, and the search strategy was limited to Chinese and English literature, potentially introducing geographical and language selection bias. This limitation implies that directly applying these models to non-Asian populations may lead to performance degradation or even clinical decision risks due to race-specific biological differences, variations in healthcare systems, or the failure of key predictors. Research confirms that ethnic and regional factors (e.g., genetic background, lifestyle) significantly influence the weighting of osteoporosis-related risk factors. A multi-center cohort study showed that East Asian women have a higher risk of hip fracture at the same bone mineral density compared to White women [35]. Another Mendelian randomization study found that although African Americans have a high prevalence of vitamin D deficiency, their incidence of osteoporotic fractures is relatively low, suggesting racial differences in the association between vitamin D and fracture risk [36]. The study by Lehmann O et al. [37] revealed that when a fracture prediction model developed based on a Swiss cohort was validated in the UK, the effectiveness of its key predictors (such as the number of falls) and the model’s performance both underwent significant changes. These differences affect not only the statistical performance of the models but also constrain their clinical applicability across diverse populations. It is important to clarify that this study focuses on risk prediction for postmenopausal osteoporosis itself, not fracture prediction. Given the current general lack of cross-ethnic validation frameworks for PMOP without fracture models, we strongly recommend exercising extreme caution when considering the application of the Asian models identified in this review to non-Asian populations.

In terms of model performance, the pooled AUC value in this study was 0.78 (95% CI: 0.73, 0.83), but significant heterogeneity (I² = 98%) was observed among the models. It is noteworthy that the sample size did not significantly affect the pooled AUC (p = 0.783). This may be because larger sample studies included more complex cases, leading to a reasonable performance decline, while smaller sample studies (EPV < 20) might have overestimated performance due to the risk of overfitting. The combined effects of small-study effects and publication bias may have led to an overestimation of the pooled AUC observed in this review and increased uncertainty in performance evaluation. Therefore, the pooled AUC should be interpreted as the average performance of the included studies rather than a generalizable single effect. Nevertheless, this meta-analysis provides a comprehensive overview of the current evidence and identifies factors that may influence performance (such as the type of predictors), thereby clarifying directions for future research. Currently, most of the included models rely on conventional clinical variables. However, with the rapid development of molecular techniques in recent years, serum biochemical markers have played an increasingly important role in the diagnosis and screening of osteoporosis. Studies have shown that changes in bone metabolism markers such as β-CrossLaps, N-MID osteocalcin, and PINP, along with their sensitivity, occur significantly earlier than changes in BMD [38, 39]. Therefore, developing prediction models that combine such bone turnover markers with BMD is a highly promising approach to enhance the accuracy of fracture-free PMOP risk prediction in the future. Nevertheless, it is important to note that the clinical translation of new predictors faces multiple obstacles, including measurement costs, accessibility, and assay turnaround time. Ultimately, prospective studies are still required to evaluate whether the application of these models genuinely improves patient outcomes or healthcare efficiency.

In terms of modeling algorithms, within the conditions of the included studies in this analysis, the performance of machine learning models did not significantly surpass that of logistic regression. This observation requires cautious interpretation, as it is influenced by several methodological limitations. First, many studies had small sample sizes and a limited number of events (EPV < 20), falling far short of the data scale required for machine learning models to fully demonstrate their advantages [40]. Second, carefully constructed features and the appropriate selection of predictors often contribute more to performance improvement than merely using complex algorithms [41]. The best-performing models (AUC > 0.9) in this analysis often integrated multiple types of important information (demographics and laboratory tests and others), which was likely the key to their high performance, rather than being solely attributable to a specific algorithm. Finally, the performance of machine learning models is highly dependent on hyperparameter tuning, which the included studies may not have adequately reported or optimized. This further complicates a fair comparison between ML and LR performance within the current analysis. Therefore, in small to medium-sized samples, logistic regression, with its simpler structure and clearer assumptions, may demonstrate more stable performance. Logistic regression holds enduring value in medical prediction modeling due to its interpretability, simplicity, and computational efficiency. Future research should more reliably evaluate the relative value of different algorithms in scenarios with sufficient data volume, complex features, and strong non-linear relationships. Model selection should be based on the specific problem, data characteristics, and the need for clinical interpretability, rather than blindly pursuing algorithmic complexity.

In this study, the predictive performance of the models was relatively good, but there were some methodological limitations in model development and validation, leading to a generally high risk of bias in the included studies. The primary causes of high risk of bias included overfitting risks, insufficient sample sizes, and inadequate handling of missing data. Among these, an insufficient sample size (EPV < 20) is the most direct cause of overfitting, which can significantly overestimate model performance metrics (such as AUC) and severely impact the model’s calibration and discriminative abilities in new populations. Regarding data sources, most of the data in this study were derived from databases or clinical records, utilizing retrospective data to construct the models. With the advancement of the information age, public databases have increasingly demonstrated their utility in the medical field due to their large sample sizes and strong statistical power. While such data are easily accessible and convenient, this design is prone to bias in establishing temporal relationships between predictors and outcomes [42], and interact with non-standardized measurements, further exacerbating model calibration errors and leading to systematic biases in risk prediction. Prospective cohort designs, randomized controlled trials, or well-designed nested case-control studies should be prioritized for future model development as they provide more robust evidence for establishing causal relationships. All studies had an events per variable (EPV) value of less than 20, likely contributing to its unstable validation results. Research shows EPV ≥ 20 for robust inference [20], underscoring the need for larger datasets in PMOP prediction. Future studies must establish stringent EPV targets a priori. When working with limited data, researchers should proactively employ feature selection techniques, parsimonious model architectures or enalized regression approaches to effectively manage model complexity while maintaining predictive validity. During the research implementation phase, some studies employed only univariate analysis to screen for predictive factors, which is extremely detrimental to the model. Not only can it lead to the omission of important predictors due to multicollinearity among independent variables, but it also significantly increases the probability of Type I errors. As a result, the final model may contain numerous spurious associations, substantially compromising its extrapolation performance [43]. Stepwise regression or penalty-/embedded-based feature selection methods demonstrate superior efficacy in handling multicollinearity and performing variable screening [44]. Deng F et al. [45] demonstrated in colorectal cancer research that an RFE-based feature selection framework incorporating multiple algorithms (logistic regression, SVM, random forest, XGBoost, and stacking) can effectively enhance classification performance when handling high-dimensional, redundant, and imbalanced data. In addition to statistical considerations, potential predictors should be comprehensively selected based on clinical relevance, measurement accessibility, and associated costs. Regarding missing data handling, three studies [25, 26, 29] failed to report missing values, while other studies employed either multiple imputation or direct elimination. Improper handling of missing data can introduce selection bias, particularly when the data is not missing completely at random. This can lead to biased estimates of model parameters. Simply deleting missing data may result in the retention of a significant number of outliers in the analytical dataset, thereby compromising predictive accuracy and calibration precision [46]. We recommend employing appropriate missing data handling methods such as multiple imputation or weighting techniques tailored to the type and mechanism, thereby minimizing potential bias. Overall, deficiencies in the analytical domain (such as insufficient sample size and overfitting) pose the most direct and severe threats to model reliability, as they directly impact the model’s internal validity. In contrast, biases in the participant and predictor domains primarily affect the model’s external validity and clinical applicability. Future research should prioritize ensuring adequate sample sizes, employing advanced statistical methods to prevent overfitting, and simultaneously minimizing foundational biases through prospective designs and standardized measurements whenever possible.

This study identified multiple risk prediction models for postmenopausal osteoporosis, yet significant publication bias was observed. Systematic omission of negative results (underperforming models) likely led to overestimated pooled AUC values. The interaction between high heterogeneity and bias substantially increased uncertainty in model performance evaluation. These findings reflected substantial methodological or population characteristic variations in current prediction model research, coupled with the preferential publication of small-sample positive results. Future studies should incorporate grey literature searches, conduct individual participant data meta-analyses to correct for bias, and prioritize clinically implemented models that have undergone rigorous external validation, thereby avoiding overreliance on potentially inflated pooled estimates. On the other hand, the model’s clinical implementation faces several challenges. First, model complexity and data requirements pose significant barriers, with some models incorporating 10–15 predictor variables and relying on laboratory indicators often unavailable in primary care settings. Second, inadequate workflow integration creates additional burdens as most models lack automated interfaces with electronic health record systems. Third, clinicians’ trust in the interpretability, reliability, and practical utility of complex models particularly “black-box” algorithms remains a pivotal determinant of adoption. Clinicians require clear understanding of a model’s decision-making logic particularly which key variables drive high-risk predictions to establish trust and guide personalized interventions. The lack of clear clinical guidance and insufficient integration of model outputs with existing diagnostic-therapeutic workflows further hinder their implementation. To facilitate clinical translation, we propose a stepped implementation strategy: (1) developing simplified tools (retaining 3–5 core clinical variables) to balance accuracy and practicality [47], ensure clear, intuitive result outputs and then seamlessly integrated into clinical decision nodes; (2) embedding models into clinical information systems or developing mobile applications for automated calculations; (3) establishing multidisciplinary teams to adapt models locally, creating tailored versions for different healthcare tiers and enhancing model interpretability by integrating explanation techniques like SHAP values and LIME or developing simplified rule sets, thereby making prediction logic transparent to address clinicians’ “black box” concerns and foster trust-building; (4) Conducting implementation research to evaluate real-world barriers, including usability, workflow impact, clinician and patient acceptance, as well as the actual effects of models on clinical outcomes such as fracture rates, treatment adherence, and cost-effectiveness [48]. Implementation studies should include dedicated evaluations of changes in physician trust levels and utilization of interpretability tools. Crucially, successful clinical integration requires concurrent healthcare provider training which key content should focus on transparently communicating the model’s principles, result interpretation guidelines, limitations, and individualized clinical integration pathways to directly address clinicians’ concerns and facilitate informed adoption. And then establishing continuous improvement mechanisms through regular outcome evaluations, ultimately enabling the transition from research tools to clinical decision support systems.

Limitation

This study has several limitations. First, significant heterogeneity existed among the included studies, primarily manifested through substantial variations in sample sizes and a predominance of retrospective designs, which may affect the generalizability of conclusions. Second, all studies originated from Asian populations, limiting the applicability of findings to other ethnic groups. Third, most studies inadequately reported calibration metrics, hindering comprehensive evaluation of model calibration performance. The literature search was restricted to studies published after 2014 and included only Chinese and English publications, which may have introduced selection bias by omitting important non-Chinese/non-English studies and pre-2014 evidence on model validation. However, given this study’s focus on machine learning models (which rapidly advanced post-2014) and Asia’s status as the most active region for osteoporosis prediction research, the core conclusions are likely less affected by this bias. Future studies should conduct multi-center prospective studies, employ standardized reporting guidelines (e.g., TRIPOD statement) and conduct model validation in more diverse populations.

Conclusion

Current machine learning-based prediction models for postmenopausal osteoporosis without fractures, as presented in the included studies, demonstrate good discriminative ability but are generally characterized by a high risk of bias, a notable lack of calibration performance evaluation, and insufficient validation of clinical utility. Furthermore, existing models are developed entirely on Asian population data, which limits their generalizability to other populations. Future research should focus on strictly adhering to prediction model research guidelines (such as PROBAST), enhancing the reporting of model calibration and clinical utility, and assessing model generalizability through external validation in multi-center studies encompassing diverse ethnicities and regions.

Supplementary Information

Supplementary Material 1. (280.6KB, docx)

Acknowledgements

Thank all participants for their kindly help during the formation and finalization processes of the manuscript.

Authors’ contributions

Data curation: Lingjia Li, Xiangzhou Lan, Yujiao Xu; Methodology and software: Weike Zeng; Supervision: Qing Chen; Visualization: Lingjia Li, Xiangzhou Lan; Writing original draft: Lingjia Li.

Funding

Project commissioned by Hunan University of Traditional Chinese Medicine Joint Public Fund Project (No: 2023XYLH008); Hunan Provincial Natural Science Foundation Program (No: 2024JJ9424); National Chinese Medicine Advantage Speciality - Nursing (State Medical Letter of Traditional Chinese Medicine [2024] No. 90);Hunan Nursing Association Training Program(HNKYP202308).

Data availability

No datasets were generated or analysed during the current study.

Declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Clynes MA, Harvey NC, Curtis EM, et al. The epidemiology of osteoporosis. Br Med Bull. 2020;133(1):105–17. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Salari N, Ghasemi H, Mohammadi L, et al. The global prevalence of osteoporosis in the world: a comprehensive systematic review and meta-analysis. J Orthop Surg Res. 2021;16(1):609. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Kherrab A, Toufik H, Ghazi M, et al. Prevalence of postmenopausal osteoporosis in morocco: a systematic review and meta-analysis. Arch Osteoporos. 2024;19(1):61. [DOI] [PubMed] [Google Scholar]
  • 4.Chai B, Feng HY, Chang Q, et al. Analysis of the prevalence of postmenopausal osteoporosis and the detection rate of bone mineral density in various regions of China. J Practical Orthop. 2020;26(09):792–6. [Google Scholar]
  • 5.Epidemiological survey of osteoporosis in. China and the results of the healthy bones special action published. Chin J Osteoporos Bone Mineral Res. 2019;12(04):317–8. [Google Scholar]
  • 6.Kanis JA. Assessment of fracture risk and its application to screening for postmenopausal osteoporosis: synopsis of a WHO report. WHO Study Group. Osteoporos Int. 1994;4(6):368–81. [DOI] [PubMed] [Google Scholar]
  • 7.Welch JM, Klifto CS, Klifto KM, et al. Prevalence and predictors of bone mineral density testing after distal radius fracture in menopausal women. Injury. 2025;56(3):112219. [DOI] [PubMed] [Google Scholar]
  • 8.Koh LK, Sedrine WB, Torralba TP, et al. A simple tool to identify Asian women at increased risk of osteoporosis. Osteoporos Int. 2001;12(8):699–705. [DOI] [PubMed] [Google Scholar]
  • 9.Cadarette SM, Jaglal SB, Kreiger N, et al. Development and validation of the osteoporosis risk assessment instrument to facilitate selection of women for bone densitometry. CMAJ. 2000;162(9):1289–94. [PMC free article] [PubMed] [Google Scholar]
  • 10.Lydick E, Cook K, Turpin J, et al. Development and validation of a simple questionnaire to facilitate identification of women likely to have low bone density. Am J Manag Care. 1998;4(1):37–48. [PubMed] [Google Scholar]
  • 11.Bui MH, Dao PT, Khuong QL, et al. Evaluation of community-based screening tools for the early screening of osteoporosis in postmenopausal Vietnamese women. PLoS One. 2022;17(4):e0266452. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Hsieh WT, Groot TM, Yen HK, et al. Validation of ten osteoporosis screening tools in rural communities of Taiwan. Calcif Tissue Int. 2024;115(5):507–15. [DOI] [PubMed] [Google Scholar]
  • 13.Ang SB, Xia JY, Cheng SJ, et al. A pilot screening study for low bone mass in Singaporean women using years since menopause and BMI. Climacteric. 2022;25(2):163–9. [DOI] [PubMed] [Google Scholar]
  • 14.Tartibian B, Fasihi L, Eslami R. Prediction of osteoporosis by K-NN algorithm and prescribing physical activity for elderly women. New Approaches in Exercise Physiology. 2020;2(4):87–100. [Google Scholar]
  • 15.Fasihi L, Tartibian B, Eslami R, et al. Artificial intelligence used to diagnose osteoporosis from risk factors in clinical data and proposing sports protocols. Sci Rep. 2022;12(1):18330. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Wu Y, Chao J, Bao M, et al. Predictive value of machine learning on fracture risk in osteoporosis: a systematic review and meta-analysis. BMJ Open. 2023;13(12):e071430. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Geersing GJ, Bouwmeester W, Zuithoff P, et al. Search filters for finding prognostic and diagnostic prediction studies in medline to enhance systematic reviews. PLoS One. 2012;7(2):e32844. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Debray TP, Damen JA, Riley RD, et al. A framework for meta-analysis of prediction model studies with binary and time-to-event outcomes. Stat Methods Med Res. 2019;28(9):2768–86. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Moons KG, de Groot JA, Bouwmeester W, et al. Critical appraisal and data extraction for systematic reviews of prediction modelling studies: the CHARMS checklist. PLoS Med. 2014;11(10):e1001744. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Peduzzi P, Concato J, Kemper E, et al. A simulation study of the number of events per variable in logistic regression analysis. J Clin Epidemiol. 1996;49(12):1373–9. [DOI] [PubMed] [Google Scholar]
  • 21.Moons KGM, Wolff RF, Riley RD, et al. PROBAST: a tool to assess risk of bias and applicability of prediction model studies: explanation and elaboration. Ann Intern Med. 2019;170(1):W1–33. [DOI] [PubMed] [Google Scholar]
  • 22.Higgins JP, Thompson SG, Deeks JJ, et al. Measuring inconsistency in meta-analyses. BMJ. 2003;327:557–60. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Shim JG, Kim DW, Ryu KH, et al. Application of machine learning approaches for osteoporosis risk prediction in postmenopausal women. Arch Osteoporos. 2020;15(1):169. [DOI] [PubMed] [Google Scholar]
  • 24.Wu Y, Chao J, Bao M, et al. Construction of predictive model for osteoporosis related factors among postmenopausal women on the basis of logistic regression and Bayesian network. Prev Med Rep. 2023;35:102378. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Wang SR, Liang SQ, Lin CZ, et al. Predictive value of nomogram model based on serum β-CTx, PINP and OC for postmenopausal osteoporosis. Chin J Woman Child Health Res. 2023;34(05):53–9. [Google Scholar]
  • 26.Liu GK, Gao Y, Shi W, et al. Correlation analysis and individualized model prediction between reproductive characteristics of postmenopausal women and osteoporosis. Chin J Osteoporos. 2023;29(03):365–70. [Google Scholar]
  • 27.Kwon Y, Lee J, Park JH, et al. Osteoporosis pre-screening using ensemble machine learning in postmenopausal Korean women. Healthcare. 2022;10(6):1107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Makond B, Pornsawad P, Thawnashom K. Decision tree modeling for osteoporosis screening in postmenopausal Thai women. Informatics. 2022;9(4):83. [Google Scholar]
  • 29.Wang JL, Pan FM, Kong C, et al. Construction and effect of a nomogram clinical prediction model for predicting osteoporosis in asymptomatic elderly women. J Capital Med Univ. 2023;44(04):629–38. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Lu XD, Wei JH, Shen JT, et al. Methods and processes for producing a systematic review of predictive model studies. Chin J evidence-based Med. 2023;23(05):602–9. [Google Scholar]
  • 31.Wynants L, Van Calster B, Collins GS, et al. Prediction models for diagnosis and prognosis of covid-19: systematic review and critical appraisal. BMJ. 2020;369:m1328. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Collins GS, Dhiman P, Andaur Navarro CL, et al. Protocol for development of a reporting guideline (TRIPOD-AI) and risk of bias tool (PROBAST-AI) for diagnostic and prognostic prediction model studies based on artificial intelligence. BMJ Open. 2021;11(7):e048008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Vickers AJ, Elkin EB. Decision curve analysis: a novel method for evaluating prediction models. Med Decis Mak. 2006;26(6):565–74. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Ramspek CL, Jager KJ, Dekker FW, et al. External validation of prognostic models: what, why, how, when and where? Clin Kidney J. 2020;14(1):49–58. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Morin SN, Berger C, Papaioannou A, et al. Race/ethnic differences in the prevalence of osteoporosis, falls and fractures: a cross-sectional analysis of the Canadian longitudinal study on aging. Osteoporos Int. 2022;33(12):2637–48. [DOI] [PubMed] [Google Scholar]
  • 36.Nethander M, Coward E, Reimann E, et al. Assessment of the genetic and clinical determinants of hip fracture risk: Genome-wide association and Mendelian randomization study. Cell Rep Med. 2022;3(10):100776. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Lehmann O, Mineeva O, Veshchezerova D, et al. Fracture risk prediction in postmenopausal women with traditional and machine learning models in a nationwide, prospective cohort study in Switzerland with validation in the UK biobank. J Bone Min Res. 2024;39(8):1103–12. [DOI] [PubMed] [Google Scholar]
  • 38.Eastell R, Szulc P. Use of bone turnover markers in postmenopausal osteoporosis. Lancet Diabetes Endocrinol. 2017;5(11):908–23. [DOI] [PubMed] [Google Scholar]
  • 39.Tartibian B, Fasihi L, Eslami R. Correlation between serum calcium, phosphorus, and alkaline phosphatase indices with lumbar bone mineral density in active and inactive postmenopausal women[J]. J Arak Univ Med Sci. 2022;25(1):120–33. [Google Scholar]
  • 40.Choi RY, Coyner AS, Kalpathy-Cramer J, et al. Introduction to machine learning, neural networks, and deep learning. Transl Vis Sci Technol. 2020;9(2):14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Christodoulou E, Ma J, Collins GS, et al. A systematic review shows no performance benefit of machine learning over logistic regression for clinical prediction models. J Clin Epidemiol. 2019;110:12–22. [DOI] [PubMed] [Google Scholar]
  • 42.Tao LY, Liu Y, Zeng L, et al. Interpretation of the reporting specification for individual -specific prognostic or diagnostic multifactorial predictive modelling (TRIPOD). Natl Med J China. 2018;98(44):3556–60. [Google Scholar]
  • 43.Wolff RF, Moons KGM, Riley RD, et al. PROBAST: a tool to assess the risk of bias and applicability of prediction model studies. Ann Intern Med. 2019;170(1):51–8. [DOI] [PubMed] [Google Scholar]
  • 44.Zhou ZR, Wang WW, Li Y, et al. In-depth mining of clinical data: the construction of clinical prediction model with R. Ann Transl Med. 2019;7(23):796. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Deng F, Zhao L, Yu N, et al. Union with recursive feature elimination: a feature selection framework to improve the classification performance of multicategory causes of death in colorectal cancer. Lab Invest. 2024;104(3):100320. [DOI] [PubMed] [Google Scholar]
  • 46.Haribhakti N, Agarwal P, Vida J, et al. A simple scoring tool to predict medical intensive care unit readmissions based on both patient and process factors. J Gen Intern Med. 2021;36(4):901–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Moons KG, Altman DG, Reitsma JB, et al. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): explanation and elaboration. Ann Intern Med. 2015;162(1):W1–73. [DOI] [PubMed] [Google Scholar]
  • 48.Proctor E, Silmere H, Raghavan R, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health. 2011;38(2):65–76. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Material 1. (280.6KB, docx)

Data Availability Statement

No datasets were generated or analysed during the current study.


Articles from BMC Musculoskeletal Disorders are provided here courtesy of BMC

RESOURCES