Skip to main content
Digital Health logoLink to Digital Health
. 2024 Mar 6;10:20552076241238093. doi: 10.1177/20552076241238093

Comparison of conventional mathematical model and machine learning model based on recent advances in mathematical models for predicting diabetic kidney disease

Yingda Sheng 1,2, Caimei Zhang 1,2, Jing Huang 1,2, Dan Wang 1,2, Qian Xiao 1,2, Haocheng Zhang 3, Xiaoqin Ha 2,
PMCID: PMC10921860  PMID: 38465295

Abstract

Previous research suggests that mathematical models could serve as valuable tools for diagnosing or predicting diseases like diabetic kidney disease, which often necessitate invasive examinations for conclusive diagnosis. In the big-data era, there are several mathematical modeling methods, but generally, two types are recognized: conventional mathematical model and machine learning model. Each modeling method has its advantages and disadvantages, but a thorough comparison of the two models is lacking. In this article, we describe and briefly compare the conventional mathematical model and machine learning model, and provide research prospects in this field.

Keywords: Mathematical model, machine learning model, diabetic kidney disease, ‌conventional model

Introduction

According to the 10th edition of the International Diabetes Federation Diabetes Atlas, 1 up to 10% of adults worldwide have been exposed to diabetes, which is linked to increasing morbidity. Diabetic kidney disease (DKD) is one of the main complications of diabetes and the main cause of end-stage renal disease.2,3 Indeed, chronic kidney diseases have been observed in 50% and 30% of patients with Type 1 diabetes and Type 2 diabetes, respectively. 4 Although early diagnosis and comprehensive treatment can prevent kidney diseases and delay their progression, 5 presently, diagnosis and rational treatment have limited clinical use in DKD patients. 6 Furthermore, the gold standard for DKD diagnosis is an invasive biopsy, which is associated with poor patient compliance. Hence, a non-invasive diagnosis of DKD is urgently needed. 7

Most studies used conventional modeling methods such as logistic regression, Cox survival regression, and multivariate analysis to statistically analyze basic information, laboratory indicators, and contributing factors to DKD, establish prediction models, and explore independent risk factors linked to DKD.834 Additionally, machine learning methods have been employed to establish risk prediction models for DKD.3543 Although different modeling methods have been used in these studies, they can be generally divided into the conventional mathematical models and the machine learning models. Each type of model has its advantages and disadvantages. The lack of comprehensive reliability evaluation prompts researchers to generally prefer conventional modeling methods, which are simpler and more user-friendly compared to machine learning methods. However, a thorough comparison of these models is lacking. Therefore, this study compares the two models to provide the best-performing model in DKD prediction.

Conventional model

In mathematical modeling, a complex problem is constructed into a mathematical problem and solved by exploring its unique internal laws and using appropriate mathematical methods to connect key factors related to the problem directly or indirectly. 44 Common mathematical models include statistical regression model, algebraic equation and difference equation model, differential equation model, discrete model, probability model, and stochastic decision analysis method. These models have been categorized as diagnostic models and prediction models according to their application value in diseases.

As previously reported, the prediction model has wider application prospects than the diagnostic model. Therefore, conventional models are established based on statistical methods for constructing prediction models. First, the data are screened, and the data meeting the criteria are selected. Next, values are assigned to the classification variables, and the data is standardized in the training and validation sets to reduce errors during validation. A single-factor analysis is then performed to identify factors with significant differences by applying statistical methods such as t-test, Kruskal-Wallis test, ANOVA, and χ2 test. Factors without differences are converted using log transformation, reciprocal conversion, or square conversion, and then analyzed using the aforementioned statistical methods to evaluate differences. The independent variables were screened by LASSO regression, elastic network graph, stepwise regression, optimal subset, and other methods. Subsequently, classical models such as the Logistic regression equation, Cox survival regression equation, unit linear regression equation, and multiple linear regression equation are used to introduce the selected independent variables and construct a diagnostic or prediction model. Finally, the proposed model is evaluated, validated, visualized, and made into application software (Figure 1).

Figure 1.

Figure 1.

Conventional model modeling process.

In the conventional model, data can be processed using simple statistical methods, and this may not affect the implementation. Because of its simplicity, conventional modeling can be readily grasped by researchers with some statistical background within a relatively short timeframe. In addition, selecting the corresponding classical model based on the research question will not affect its prediction efficiency. Although the conventional model is easy and has good prediction performance, it also has some shortcomings. It can only perform optimally when the sample size is neither too large nor too small. A larger sample size of the conventional model will lead to a labor-intensive process of selecting data that meets the requirements, with a high risk of error and bias. On the other hand, a smaller sample size of the conventional model may lead to inaccurate results when applying statistical methods. Because the formula of the classical model is fixed, the range of problems that can be studied using this method is limited. For instance, if the problem of interest does not match existing classical models, applying a conventional model may lead to inaccurate results. Generally, the conventional model includes only those influencing factors screened out as significantly different using statistical analysis and discards those without differences. However, some non-differential factors or those with combined effects with other factors may be excluded, which may lead to incomplete results, and even inaccurate or unconvincing model predictions (Table 1).

Table 1.

Advantages and disadvantages of the models.

Model Advantages Disadvantages
Conventional model Simplicity Time-consuming and laborious
Convenient Average generalization ability
Low equipment requirements Small sample
Mature operation process Possible omissions in factor selection
Good predictive performance (no need for parameter adjustment) Prone to deviations, underfitting, or overfitting
Few types of models
Machine Learning Model Labor-saving Complex
Large-sample High equipment requirements
Comprehensive selection of factors Difficult to operate and difficult to learn
Reduce the occurrence of deviations, underfitting, and overfitting(require cross-validation, parameter tuning, and other processes) Require cross-validation, parameter tuning, and other processes to improve model performance
Excellent predictive performance (require cross-validation, parameter tuning, and other processes) The operation process is not very mature, there is still room for exploration.
Excellent generalization ability (require cross-validation, parameter tuning, and other processes) Establishing a model requires making multiple models and comparing them to select one with good performance
Multiple types of algorithms Poor interpretability (the general ranking of ability of interpretation: regression >cluster classification> tree-based model > Neural Network)

Recently, artificial intelligence (AI) has attracted great attention, and some researchers have even attempted to apply AI algorithms to establish better mathematical models than conventional models. Nevertheless, these novel models and conventional models have not been compared in performance.

Machine learning

AI is the hallmark of the progress in computing from “data processing” to “knowledge processing.” With the rapid advances in computer technology, the question arises whether AI can replace humans in intellectual activities. 45 AI research seeks to create computers that can emulate the intelligent actions of humans in their environment. Machine learning, as a branch of AI, serves as a fundamental tool for constructing prediction models.

Machine learning is an interdisciplinary field that integrates probability theory, statistics, approximation theory, convex analysis, and algorithmic complexity to study how computers simulate or implement human learning behaviors. Machine learning focuses on harnessing computational power and human knowledge to optimize system performance. In computer systems, “experience” commonly exists as “data.” Hence, machine learning research mainly focuses on creating a “learning algorithm” that could generate “models” from data archived in a computer. A learning algorithm can generate models based on the empirical data provided. In a new situation, the model will provide us with corresponding judgments. If computer science is the study of “algorithms,” then machine learning is the study of “learning algorithms.” 46 Machine learning is at the core of AI, and it is fundamental to creating intelligent computers. 47 Common models used for data mining and statistical machine learning include linear regression and linear classification algorithms and model evaluation and selection algorithms, such as decision tree and combination method, support vector machine, neural network, and deep learning methods.4850 Initially, we conduct data cleaning, noise reduction, missing value processing, feature engineering, normalization, and standardization. Next, a suitable model is selected for training based on data characteristics, model use, experimental purpose, etc. (multiple models can be selected for comparison and the best selection). Once the model is trained, the performance of the model is evaluated and verified. In cases with an insufficient sample size or there is no independent verification set, we perform cross-verification to prevent problems such as overfitting of the model. The commonly used cross-verification methods include: HoldOut cross-validation, K-fold cross-validation, hierarchical K-fold cross-validation, leave P out cross-validation, leave a cross-validation, shuffle-split, rolling cross-validation (time series), etc. Among them, the rolling cross-validation is considered the most effective, but it cannot be used for random samples. The order of the data is important. Cross-validation employs various strategies to split the main dataset into training and validation sets, each with its unique approach. For instance, K-fold cross-validation partitions the data into K-equal subsets, utilizing one-fold for validation and K−1 folds for training. Rolling cross-validation, on the other hand, divides the data into training and validation sets based on temporal order. Finally, to improve the performance of the model and achieve its generalization performance on the verification set or test set (reduce the problem of underfitting or overfitting and error of the model), the parameters of the model should be adjusted (Figure 2).

Figure 2.

Figure 2.

Machine learning model modeling process.

There are two types of parameters in machine learning, one is the parameter, and the other is the hyperparameter. Since the parameter is determined using the model selection algorithm, tuning refers to the hyperparameter. Two approaches are used to conduct parameter adjustment, one is manual parameter adjustment, and the other is automatic parameter adjustment (application algorithm parameter adjustment). Manual parameter tuning, being a time-consuming and laborious process, defeats the purpose of employing machine learning models to enhance efficiency. Consequently, automatic parameter adjustment is adopted as a more efficient alternative. Common auto-tuning methods include grid search, random search, Bayesian optimization, ensemble learning, and others. The choice of parameter tuning method depends on the specific machine learning model selected. For instance, Bayesian optimization is employed to adjust models derived from Bayesian inference. Furthermore, cross-validation remains an essential component of the parameter adjustment process. For example, K-fold cross-validation is often applied to reduce the estimation bias and variance caused by the randomness of data division, because the adjustment of hyperparameters also needs to take into account the structural risk and empirical risk of the model. The parameter adjustment is completed when the model evaluation index reaches the optimum. For example, precision, accuracy, recall, F1 score, and area under the curve of the receiver operating characteristic (AUC-ROC) are all the largest or the larger the better; mean square error, root mean square error, mean absolute error, etc. reach the minimum value or the smaller the better. So based on the characteristics of the above indicators, the performance of different models can be compared by comparing these indicators of different models.

Machine learning can solve problems with larger sample sizes and reduce errors in workforce statistics and data screening. However, if the included data deviates significantly from the real problem, there may be a bias. Machine learning provides several learning algorithms and can create a novel learning algorithm to address a specific problem. Therefore, machine learning can explore a much wider range of problems than the conventional model, but this may result in over-fitting, which, like under-fitting, is not suitable for model application. Due to the overlapping capabilities of various learning algorithms, certain problems can be addressed using multiple algorithms, although with varying prediction accuracies. Thus, it is necessary to establish and compare models to select models optimal for a particular problem separately. For data analysis, machine learning does not randomly discard data. However, some factors that are not significant or useful for the problem may be included in the model, thereby reducing the predictive effectiveness. Because of this characteristic, machine learning will perfectly preserve the undifferentiated but related data that are filtered out when applying the conventional model. While machine learning models can suffer from errors like deviations, overfitting, and underfitting, these issues can be significantly mitigated through techniques like cross-validation and parameter tuning. Since the working principles behind machine learning are not very clear, its interpretability is relatively low. The general ranking of the ability of interpretation: regression > cluster classification > tree-based model > neural network. Despite their relative complexity and high operating coefficients, machine learning models remain highly sought-after due to their remarkable predictive capabilities (Table 1).

Discussion

Two common modeling approaches were compared for predicting DKD risk (Table 2): the conventional mathematical model and the machine learning model, including sample size, key features, performance of different algorithms (such as AUC-ROC), and machine learning also described whether parameters were adjusted and whether cross-validation was performed.

Table 2.

Comparison of performance of various models.

Author Model Model specifications Sample size(N Key features Model performances Cross-validation Tuning parameters
AUC-ROC Others
Tan et al. 10 Conventional Logistic regression and nomogram N:102 Duration of T2DM, HbA1c, presence of DR, absence of hematuria, and absence of systemic biomarkers. 88.60% Positive predictive value (PPV):87.5%, specifcity:81.1%, negative predictive value (NPV):65.2%, sensitivity:75.4% Internal validation with bootstrap sampling with 200
Lv et al. 11 Conventional Cox regression model N:515 Triglyceride glucose (TyG) index 69.00% Sensitivity:71.3% and specificity:37.2%
Sanchez-Alamo et al. 12 Conventional Time-to-event curves were based on the Kaplan–Meier analysis and Cox regression model N:103 Interleukin-6 (IL-6) 87.90% Cut-off:4.68 pg/mL, sensitivity:100%, specificity:78.72%, PPV: 45.6%, and NPV: 100%
Yang and Jiang 19 Conventional Logistic regression, nomogram, and forest plots N:706 Scr, hypertension, HbA1c, BUN, BMI, TG, and DPN 77.30% C-index:0.773
Hu et al. 25 Conventional LASSO regression, logistic regression, and nomogram N:3489 SBP, DBP, FBG, HbA1c, TGs, SCR, BUN, and BMI 74.40% C-index:0.744
Hosseini Sarkhosh et al. 35 Machine learning Random forest (RF) and logistic regression (LR) Training set:1907, verification set:1543, N:3450 diabetes duration, HbA1c, eGFR, ACR,
CVD, and hypertension
75.50% Accuracy:0.7239, precision:0.6908, and F1 score:0.6051 Recursive feature elimination with cross-validation (RFECV) Grid search technique
Zou et al. 36 Machine learning and conventional RF and nomogram N:390 Cys-C, sAlb, Hb, UTP, and eGFR 90% Accuracy:82.65%, sensitivity:83.33%, and specificity:81.58% 10-fold cross-validation
Zou et al. 36 Machine learning Gradient boosting machine (GBM) N:390 Cys-C, sAlb, Hb, UTP, and eGFR 88% Accuracy:83.67%, sensitivity:95%, and specificity:65.79% 10-fold cross-validation
Zou et al. 36 Machine learning Support vector machine (SVM) N:390 Cys-C, sAlb, Hb, UTP, and eGFR 88% Accuracy:83.67%, sensitivity:86.67%, and specificity:78.95% 10-fold cross-validation
Zou et al. 36 Machine learning LR N:390 Cys-C, sAlb, Hb, UTP, and eGFR 83% Accuracy:79.59%, sensitivity:78.33%, and specificity:81.58% 10-fold cross-validation
Zhang et al. 37 Machine learning RF N:929 DR, DM course, Hb, PP, sCr, ALB,
TC, sudden onset of heavy proteinuria, hematuria, and family history
of DM
95.30% Accuracy:88%, sensitivity:84.4%, specificity:89.9%, and balanced accuracy:87.1% Five-fold cross-validation: 1. AUC-ROC:94.6%,
2. AUC-ROC:94.6%,
3. AUC-ROC:97.4%,
4. AUC-ROC:93.8%,
5. AUC-ROC:96%
Zhang et al. 37 Machine learning SVM N:930 DR, DM course, Hb, PP, sCr, ALB, TC,
sudden onset of heavy proteinuria, hematuria, family history of DM
94.70% Accuracy:88.3%, sensitivity:84.2%, specificity:90.6%, and balanced accuracy:87.4% Five-fold cross-validation: 1. AUC-ROC:94.8%,
2. AUC-ROC:92.8%,
3. AUC-ROC:97.2%,
4. AUC-ROC:94%,
5. AUC-ROC:94.7%
Allen et al. 38 Machine learning RF/gradient boosted trees (XGB) N:111046 82.3%/82.5% Sensitivity: 75%/75%, Specificity:73.9%/74.2%, Yes
Liu et al. 40 Machine learning Bayesian network (BN) N:1485 83.10% 10-fold cross-validation Tabu search

In general, the sample size of machine learning is larger than that of the traditional model. Based on AUC-ROC, the performance of the machine learning model was higher than that of the traditional model. Most machine learning models use cross-validation. Some studies have applied tuning parameters, such as Zhang et al., 35 Allen et al., 38 and Liu et al. 40 In part, based on the AUC-ROC values including those reported by Tanet al., 10 Sanchez-Alamo et al., 12 Zhang et al., 35 and Allenet al., 38 the performance of the traditional model is not necessarily lower than that of machine learning. Through the adoption of model optimization processes such as cross-validation and parameter adjustment, the performance of machine-learning models can be greatly improved. A pending question is, can these methods of mathematical model optimization improve the performance of traditional models if they are applied to traditional models?

In this article, we only made superficial comparisons between traditional models and machine learning models based on previous findings. While both models possess their own unique strengths and limitations, further comprehensive research is needed to directly compare their performances under identical conditions. Consequently, definitive conclusions regarding the superiority of either model remain elusive. In the era of big data, machine learning models may have more development prospects and application value. However, whether conventional mathematical models should be completely abandoned is a question that cannot be definitively answered.

The use of prediction models has become increasingly prevalent and influential in the medical field. However, current prediction models show different test power. To integrate research and clinical medicine to develop a reliable mathematical model with high inspection efficiency, we should compare the performance of tools for model establishment under different conditions.

Several research directions are proposed. First, under the premise of unified standards, the same research questions, data set, quantity, and influencing factors should be used. In addition, appropriate adjustments should be made, including selecting multiple types of questions, heterogeneous of population, time factors (e.g. chronology, survival, mortality, etc.), multiple outcomes, using different sample sizes (small, medium, or large), and exploring more influencing factors matching the research questions. The optimized model or algorithm should be selected for modeling using the two types of modeling methods: conventional modeling and machine learning. Finally, the two methods can be compared to identify the more effective one. Alternatively, according to the research method of Zou et al., 36 the combination of these two methods may build a better prediction model, and even a prediction power of 1 + 1 > 2 may appear. The author has carried out some related experiments and obtained some results, among which the combined model is better than the single algorithm model for the same research problem under the same small sample data set.

Conclusion

In summary, this paper provides a comparison between conventional models and machine learning models, highlighting their respective advantages and disadvantages. However, it is worth noting that the existing research lacks a detailed comparative analysis at an equivalent level. Furthermore, our findings indicate that the fusion model can exhibit superior prediction performance compared to both individual models. However, studies on the fusion model remain limited in number. Consequently, these aspects will be the primary focus of our future research endeavors. Currently, our team has initiated research in these areas and has achieved significant progress.

Acknowledgements

Not applicable.

Footnotes

Contributorship: YS: conceptualization, writing–original draft, and investigation; CZ: writing–original draft and investigation; JH: investigation; DW: investigation; QX: investigation; HZ: investigation; XH: supervision, writing–review, and editing, corresponding. YS and XH: I am willing to take full responsibility for the article, including the accuracy and appropriateness of the reference list. All authors have read and agreed to the published version of the manuscript.

Consent for publication: All authors are aware of and agree to publish.

The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Ethical approval and consent to participate: Not applicable.

Funding: The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This project is funded by the Fund of Gansu Provincial Health Commission of China [Approval No.: GSWSKY2022-03]. The funder does not play any role in research design, data collection and analysis, publishing decisions, or manuscript preparation.

Guarantor: YS and XH.

References

  • 1.International Diabetes Federation. IDF Diabetes Atlas, 10th ed. Brussels, Belgium: 2021. Available at: https://www.diabetesatlas.org.
  • 2.Yamazaki T, Mimura I, Tanaka Tet al. et al. Treatment of diabetic kidney disease: current and future. Diabetes Metab J 2021; 45: 11–26. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Pelle MC, Provenzano M, Busutti M, et al. Up-date on diabetic nephropathy. Life (Basel) 2022; 12: 1202. Published 2022 Aug 8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Caramori ML, Rossing P. Diabetic kidney disease. [Updated 2022 Aug 3]. In: Feingold KR, Anawalt B, Blackman MR, et al. (eds) Endotext [Internet]. South Dartmouth (MA): MDText.com, Inc.; 2000-. Available from: https://www.ncbi.nlm.nih.gov/books/NBK279103/ [Google Scholar]
  • 5.Standards of Medical Care in Diabetes - 2020. 11. Microvascular complications and foot care: standards of medical care in diabetes-2020. Diabetes Care 2020; 43: S135–S151. [DOI] [PubMed] [Google Scholar]
  • 6.Plantinga L, Boulware L, Coresh J, et al. Patient awareness of chronic kidney disease: trends and predictors. Arch Internal Med 2008; 168: 2268–2275. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Pu X. A preliminary study of functional magnetic resonance imaging on early renal function in diabetic patients. Shanxi Medical University. 2022.
  • 8.Saito H, Tanabe H, Kudo A, et al. High FIB4 index is an independent risk factor of diabetic kidney disease in type 2 diabetes. Sci Rep 2021; 11: 11753. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Zou Y, Zhao L, Zhang J, et al. Metabolic-associated fatty liver disease increases the risk of end-stage renal disease in patients with biopsy-confirmed diabetic nephropathy: a propensity-matched cohort study. Acta Diabetol 2023; 60: 225–233. [DOI] [PubMed] [Google Scholar]
  • 10.Tan HZ, Choo JCJ, Fook-Chong S, et al. Development and validation of a novel nomogram to predict diabetic kidney disease in patients with type 2 diabetic mellitus and proteinuric kidney disease. Int Urol Nephrol 2023; 55: 191–200. [DOI] [PubMed] [Google Scholar]
  • 11.Lv L, Zhou Y, Chen X, et al. Relationship between the TyG index and diabetic kidney disease in patients with type-2 diabetes mellitus. Diabetes Metab Syndr Obes 2021; 14: 3299–3306. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Sanchez-Alamo B, Shabaka A, Cachofeiro V, et al. Serum interleukin-6 levels predict kidney disease progression in diabetic nephropathy. Clin Nephrol 2022; 97: 1–9. [DOI] [PubMed] [Google Scholar]
  • 13.Koska J, Gerstein HC, Beisswenger PJ, et al. Advanced glycation end products predict loss of renal function and high-risk chronic kidney disease in type 2 diabetes. Diabetes Care 2022; 45: 684–691. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Waijer SW, Sen T, Arnott C, et al. Association between TNF receptors and KIM-1 with kidney outcomes in early-stage diabetic kidney disease. Clin J Am Soc Nephrol 2022; 17: 251–259. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Zhu H, Bai M, Xie X, et al. Impaired amino acid metabolism and its correlation with diabetic kidney disease progression in type 2 diabetes mellitus. Nutrients 2022; 14: 3345. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Sun L, Wu Y, Hua RX, et al. Prediction models for risk of diabetic kidney disease in Chinese patients with type 2 diabetes mellitus. Ren Fail 2022; 44: 1454–1461. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Ma Y, Wang Q, Chen Y, et al. Correlation of dehydroepiandrosterone with diabetic nephropathy and its clinical value in early detection. J Diabetes Investig 2022; 13: 1695–1702. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Wang YH, Chang DY, Zhao MH, et al. Glutathione peroxidase 4 is a predictor of diabetic kidney disease progression in type 2 diabetes mellitus. Oxid Med Cell Longev 2022; 2022: 2948248. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Yang J, Jiang S. Development and validation of a model that predicts the risk of diabetic nephropathy in type 2 diabetes mellitus patients: a cross-sectional study. Int J Gen Med 2022; 15: 5089–5101. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Fang WC, Chou KM, Sun CY, et al. Thermal perception abnormalities can predict diabetic kidney disease in type 2 diabetes mellitus patients. Kidney Blood Press Res 2020; 45: 926–938. [DOI] [PubMed] [Google Scholar]
  • 21.Wu Y, Liu F. Aspartate aminotransferase to alanine aminotransferase ratio and the risk of diabetic nephropathy progression in patients with type 2 diabetes mellitus: a biopsy-based study. J Diabetes Complications 2022; 36: 108235. [DOI] [PubMed] [Google Scholar]
  • 22.Li MR, Sun ZJ, Chang DY, et al. C3c deposition predicts worse renal outcomes in patients with biopsy-proven diabetic kidney disease in type 2 diabetes mellitus. J Diabetes 2022; 14: 291–297. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Zhang J, Wang Y, Zhang R, et al. Serum fibrinogen predicts diabetic ESRD in patients with type 2 diabetes mellitus. Diabetes Res Clin Pract 2018; 141: 1–9. [DOI] [PubMed] [Google Scholar]
  • 24.Satirapoj B, Pooluea P, Nata N, et al. Urinary biomarkers of tubular injury to predict renal progression and end stage renal disease in type 2 diabetes mellitus with advanced nephropathy: a prospective cohort study. J Diabetes Complications 2019; 33: 675–681. [DOI] [PubMed] [Google Scholar]
  • 25.Hu Y, Shi R, Mo R, et al. Nomogram for the prediction of diabetic nephropathy risk among patients with type 2 diabetes mellitus based on a questionnaire and biochemical indicators: a retrospective study. Aging (Albany NY) 2020; 12: 10317–10336. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Wang K, Xu W, Zha B, et al. Fibrinogen to albumin ratio as an independent risk factor for type 2 diabetic kidney disease. Diabetes Metab Syndr Obes 2021; 14: 4557–4567. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Yang Z, Duan P, Li W, et al. The correlation between thyroid hormone levels and the kidney disease progression risk in patients with type 2 diabetes. Diabetes Metab Syndr Obes 2022; 15: 59–67. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Zhou T, Shen L, Li Z, et al. Severe 25-hydroxyvitamin d deficiency may predict poor renal outcomes in patients with biopsy-proven diabetic nephropathy. Front Endocrinol (Lausanne) 2022; 13: 871571. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Gao YM, Feng ST, Yang Y, et al. Development and external validation of a nomogram and a risk table for prediction of type 2 diabetic kidney disease progression based on a retrospective cohort study in China. Diabetes Metab Syndr Obes 2022; 15: 799–811. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Chang LH, Chu CH, Huang CC, et al. Fibroblast growth factor 21 levels exhibit the association with renal outcomes in subjects with type 2 diabetes mellitus. Front Endocrinol (Lausanne) 2022; 13: 846018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Fei X, Xing M, Wo M, et al. Thyroid stimulating hormone and free triiodothyronine are valuable predictors for diabetic nephropathy in patient with type 2 diabetes mellitus. Ann Transl Med 2018; 6: 305. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Sun Z, Wang K, Miller JD, et al. External validation of the risk prediction model for early diabetic kidney disease in Taiwan population: a retrospective cohort study. BMJ Open 2022; 12: e059139. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Zhao L, Zhang J, Lei S, et al. Combining glomerular basement membrane and tubular basement membrane assessment improves the prediction of diabetic end-stage renal disease. J Diabetes 2021; 13: 572–584. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Ham YR, Lee EJ, Kim HR, et al. Ultrasound renal score to predict the renal disease prognosis in patients with diabetic kidney disease: an investigative study. Diagnostics (Basel) 2023; 13: 515. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Hosseini Sarkhosh SM, Hemmatabadi M, Esteghamati A. Development and validation of a risk score for diabetic kidney disease prediction in type 2 diabetes patients: a machine learning approach. J Endocrinol Invest 2023; 46: 415–423. [DOI] [PubMed] [Google Scholar]
  • 36.Zou Y, Zhao L, Zhang J, et al. Development and internal validation of machine learning algorithms for end-stage renal disease risk prediction model of people with type 2 diabetes mellitus and diabetic kidney disease. Ren Fail 2022; 44: 562–570. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Zhang W, Liu X, Dong Z, et al. New diagnostic model for the differentiation of diabetic nephropathy from non-diabetic nephropathy in Chinese patients. Front Endocrinol (Lausanne) 2022; 13: 913021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Allen A, Iqbal Z, Green-Saxena A, et al. Prediction of diabetic kidney disease with machine learning algorithms, upon the initial diagnosis of type 2 diabetes mellitus. BMJ Open Diabetes Res Care 2022; 10: e002560. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Fan Y, Long E, Cai L, et al. Machine learning approaches to predict risks of diabetic complications and poor glycemic control in nonadherent type 2 diabetes. Front Pharmacol 2021; 12: 665951. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Liu S, Zhang R, Shang X, et al. Analysis for warning factors of type 2 diabetes mellitus complications with Markov blanket based on a Bayesian network model. Comput Methods Programs Biomed 2020; 188: 105302. [DOI] [PubMed] [Google Scholar]
  • 41.Belur Nagaraj S, Pena MJ, Ju W, et al. Machine-learning-based early prediction of end-stage renal disease in patients with diabetic kidney disease using clinical trials data. Diabetes Obes Metab 2020; 22: 2479–2486. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Connolly P, Stapleton S, Mosoyan G, et al. Analytical validation of a multi-biomarker algorithmic test for prediction of progressive kidney function decline in patients with early-stage kidney disease. Clin Proteomics 2021; 18: 26. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Song X, Waitman LR, Yu AS, et al. Longitudinal risk prediction of chronic kidney disease in diabetic patients using a temporal-enhanced gradient boosting machine: retrospective cohort study. JMIR Med Inform 2020; 8: e15510. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Giordano FR, Fox WP, Horton SB. A First Course in Mathematical Modeling. 5th ed. US: Brooks Cole, 2013, ISBN: 9781285050904. [Google Scholar]
  • 45.Gang Z. What is artificial intelligence. China Public Science 2018; 1: 44–45. [Google Scholar]
  • 46.Zhihua Z, Ed. Machine learning. Beijing: Tsinghua Press; 2016. ISBN: 9787302423287 [Google Scholar]
  • 47.Li B, Ed. Practical role of machine learning. Beijing: Posts and Telecommunications Press, 2017. ISBN: 9787115460417 [Google Scholar]
  • 48.Xiaoling L, Jie S, Eds. Big data mining and statistical machine learning. Beijing: China Renmin University Press: Big Data Analysis and Statistics Application Series, 2016) ISBN: 9787300231013 [Google Scholar]
  • 49.Deo RC. Machine learning in medicine. Circulation 2015; 132: 1920–1930. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Choi RY, Coyner AS, Kalpathy-Cramer J, et al. Introduction to machine learning, neural networks, and deep learning. Transl Vis Sci Technol 2020; 9: 14. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Digital Health are provided here courtesy of SAGE Publications

RESOURCES