Skip to main content
PLOS One logoLink to PLOS One
. 2023 Dec 1;18(12):e0295234. doi: 10.1371/journal.pone.0295234

Chronic kidney disease prediction using boosting techniques based on clinical parameters

Shahid Mohammad Ganie 1, Pijush Kanti Dutta Pramanik 2, Saurav Mallik 3, Zhongming Zhao 4,*
Editor: Anwar PP Abdul Majeed5
PMCID: PMC10691694  PMID: 38039306

Abstract

Chronic kidney disease (CKD) has become a major global health crisis, causing millions of yearly deaths. Predicting the possibility of a person being affected by the disease will allow timely diagnosis and precautionary measures leading to preventive strategies for health. Machine learning techniques have been popularly applied in various disease diagnoses and predictions. Ensemble learning approaches have become useful for predicting many complex diseases. In this paper, we utilise the boosting method, one of the popular ensemble learnings, to achieve a higher prediction accuracy for CKD. Five boosting algorithms are employed: XGBoost, CatBoost, LightGBM, AdaBoost, and gradient boosting. We experimented with the CKD data set from the UCI machine learning repository. Various preprocessing steps are employed to achieve better prediction performance, along with suitable hyperparameter tuning and feature selection. We assessed the degree of importance of each feature in the dataset leading to CKD. The performance of each model was evaluated with accuracy, precision, recall, F1-score, Area under the curve-receiving operator characteristic (AUC-ROC), and runtime. AdaBoost was found to have the overall best performance among the five algorithms, scoring the highest in almost all the performance measures. It attained 100% and 98.47% accuracy for training and testing sets. This model also exhibited better precision, recall, and AUC-ROC curve performance.

1. Introduction

Chronic kidney disease (CKD) has become very common across races [1], resulting in millions of deaths worldwide annually [2]. Proper diagnosis and timely treatment are major concerns in most developing countries. CKD mostly hits older people [3, 4], and by 2050, the number of people aged 65 years and above is estimated to increase to 1.5 billion from 703 million in 2019, with a more than double growth rate [5]. This will put a significant additional burden on healthcare services across the countries [6].

According to a study by the Center for Disease Control and Prevention, in 2017, approximately thirty million people in the U.S. alone were affected by CKD [7], which has been increased to 37 million in 2021 [8]. Moreover, most people are not aware of being infected by CKD. Traditionally, doctors confirm the CKD for any patient based on some clinical tests such as estimating glomerular filtration rate (GFR) from a filtration marker (e.g., serum creatinine or cystatin C) or through a urine test, detecting the presence of albumin and/or protein [911]. However, these tests may not always give accurate results, leading to the wrong diagnosis.

CKD can be mitigated to some extent if the possibility of it can be predicted beforehand for the suspected patients [12, 13]. This would allow healthcare professionals to deliver better services by embracing precautionary measures and early diagnosis and treatment. Machine learning algorithms have been popularly used in several disease diagnoses and predictions [1417]. For CKD prediction also, various such techniques have been explored [1822]. Machine learning algorithms are powerful for analysing large and complex datasets and identifying patterns and relationships that may not be apparent to human experts. In the context of CKD prediction, machine learning has the potential to improve accuracy and reduce costs by identifying early signs of disease progression and predicting the risk of developing CKD in at-risk populations.

However, traditional machine learning techniques suffer from some crucial limitations, including [23, 24]:

  • Overfitting, where the algorithm becomes too specialised to the training data and fails to generalise to new data.

  • Large, high-quality datasets are needed to train and validate the algorithms, which can be challenging to obtain in some clinical settings.

  • Training and evaluating machine learning algorithms may require considerable computational time and resources, especially for large datasets.

  • High dependency on the quality and quantity of data available for training. If the data is incomplete, biased, or otherwise of poor quality, the resulting algorithm will be inaccurate or may not work at all.

  • The machine learning algorithms can inadvertently incorporate biases present in the training data, leading to unfair or discriminatory outcomes.

Recently, ensemble learning techniques have shown great promise in improving the accuracy, robustness, and generalizability of predictive models, making them valuable in many fields, including healthcare, finance, marketing, social media analytics, etc. The ensemble learning approaches are gaining attention for disease prediction with higher accuracy [2531]. Among the ensemble learning techniques such as boosting, bagging, and stacking, boosting algorithms can reduce the training error (bias) and testing error (variance).

In this paper, we design a novel CKD prediction model using boosting algorithms. We aim to improve the performance of the disease prediction model over similar existing works. The contributions of this paper are summarised as follows.

  • Exploratory data analysis is performed to transform the considered dataset for better experimental usability.

  • Hyperparameter techniques, such as standardisation, normalisation, feature selection, and fine-tuning, are employed to achieve optimal results.

  • The attribution of existing dataset features to disease prediction is assessed.

  • Five boosting algorithms are individually applied to build the prediction model.

  • The prediction performances of the five boosting algorithms are evaluated and compared.

  • Our model achieved better accuracy and runtime than other machine learning-based CKD prediction models in method evaluation.

2. Related work

As mentioned above, machine learning has been extensively used for various disease diagnoses and predictions [17, 32, 33]. To improve the performance of these models, several machine learning techniques are combined to extract the advantages of each of them. This ensemble approach has gained acceptance and popularity after successful implementations for the prediction, detection, diagnosis, and prognosis of different diseases, such as heart disease [34, 35], breast cancer [36], skin disease [37], thyroid disease [38], myocardial infarction [39], Alzheimer’s disease [40], etc. For CKD prediction, several prediction techniques and models have already been proposed [41]. In the following, we briefly review some notable experiments for the diagnosis and prediction of CKD using ensemble learning techniques.

For CKD prediction, Kumar et al. [42] proposed an ensemble learning approach that comprises a support vector machine (SVM), decision tree, C4.5 decision tree, particle swarm optimisation ‐ multilayer perceptron (PSO-MLP), and artificial bee colony C4.5. The prediction process has two steps–i) in the first step, weak decision tree classifiers are obtained from C4.5, and ii) in the second step, the weak classifiers are combined with the weighted sum to get the final output from the classifier, attaining accuracy of 92.76%. Pal [43] developed a bagging ensemble method comprising a decision tree, SVM, and logistic regression to predict CKD. The best accuracy of 95.92% was achieved in the case of the decision tree. Hasan and Hasan [44] proposed an ensemble method for kidney disease diagnosis. They used adaptive boosting (AdaBoost), bootstrap aggregating, extra trees, gradient boosting, and random forest to build their prediction model. They performed tenfold cross-validation to validate the results. The highest accuracy of 99% was attained with adaptive boosting. For CKD detection, Wibawa et al. [45] developed an ensemble learning method that comprises three stages. In the first stage, base classifiers like Naive Bayes, SVM, and k nearest neighbour (kNN) were used. Correlation-based feature selection (CFS) was combined with the base classifiers mentioned above in the second stage. In the third stage, they used CFS with AdaBoost, achieving the highest accuracy of 98.01%. For CKD diagnosis, Jongbo et al. [1] built an ensemble learning model through bagging and random subspace based on three base classifiers–kNN, naïve Bayes, and decision tree. Data preprocessing was done to mitigate the missing value issue and data normalisation for scaling the independent variables within a certain range. The random subspace gained better performance than bagging in most performance measure metrics. It achieved an accuracy of 98.30% when combined with the decision tree method. To detect CKD, Ebiaredoh-Mienye et al. [46] combined the information-gain-based feature selection technique with the proposed cost-sensitive AdaBoost (C.S. AdaBoost), intending to save CKD screening time and cost. They trained the proposed C.S. AdaBoost with the reduced feature set, which attained a maximum accuracy of 99.8%. Emon et al. [47] used various boosting techniques to predict the risk of CKD progression among patients. The authors applied the principal component analysis (PCA) method to get the optimal feature set and attained the highest accuracy rate of 99.0% using random forest (R.F.). Ramaswamyreddy et al. [48] used wrapper methods along with bagging and boosting models to develop a CKD prediction model, attaining an accuracy of 99.0% with gradient boosting. However, the authors did not evaluate their model using other performance measure metrics.

3. Research methodology

This section briefly discusses the research steps followed and the ensemble learning techniques used in the experiment.

3.1 Research workflow

The workflow of the proposed work is shown in Fig 1. We performed exploratory data analysis on the considered dataset for better quality assessment. In this phase, missing values are identified and replaced using data imputation methods. The interquartile range (IQR) method is used to detect outliers present in the dataset. Some other required libraries are executed to check the corrupt data, if any, in the dataset. Also, standardisation, normalisation, feature selection, and tuning are made during the prediction model development process using five boosting algorithms. The dataset was split into training (60%) and test (40%) subsets. The results are assessed through various performance metrics.

Fig 1. The workflow of the proposed ensemble learning based CKD prediction.

Fig 1

3.2 Boosting algorithms

Ensemble learning is a method that combines different traditional machine learning approaches to enhance the performance of the prediction model [49]. Various ensemble learning approaches are proposed [50, 51]. Boosting algorithm is one of the effective approaches in the ensemble learning family. In the literature, several boosting algorithms can be found [52, 53]. In this experiment, specifically for CKD prediction, we considered the following five ensemble learning based boosting algorithms:

XGBoost

XGBoost (eXtreme gradient boosting) works by combining different kinds of decision trees (weak learners) to calculate the similarity scores independently [54]. It helps to overcome the problem of overfitting during the training phase by adapting the gradient descent and regularisation process. The mathematical formula for the XGBoost algorithm is shown in Eq 1.

fθ(x)=m=1Tγmhmx;θm=m=1TγmlxRjm (1)

where fθ(x) is XGBoost model with parameters θ,hm is the mth weak decision tree with parameters θm, and γm is the weight associated with mth tree. T denotes the number of decision trees, l denotes the loss function, and Rjm is an indicator function that returns 1 if x is in region Rjm, otherwise 0.

CatBoost

CatBoost (categorical boosting) is faster than other boosting algorithms as it does not require the exploration of data preprocessing [55]. It is used to deal with high cardinality categorical variables. For low cardinality variables, one-hot encoding techniques are used for conversion. The objective function for the CatBoost algorithm is defined using Eq 2.

L(y,f(x))=i=1Nlyi,fxi+λ2j=1Pwj2 (2)

where y is the true label of the training set, f(x) is the predicted label, N is the number of training samples, l denotes the loss function, λ is the regularisation parameter used to penalise overfitting, P is the number of features and w is the weight associated with each feature of the dataset.

LightGBM

LightGBM is an extension of a gradient boosting algorithm, capable of handling large datasets with less memory utilisation during the model evaluation process [56]. Gradient-based one-side sampling method is used for splitting the data samples, reducing the number of features in sparse datasets during training. The objective function for the LightGBM algorithm is defined using Eq 3.

L(θ)=i=1Nlyi,y^l+i=1Tωfj (3)

where θ is a set of model parameters, N is the number of training samples, l denotes the loss function, yi is the true label of ith sample, y^l is the predicted label for the model, fj is the jth decision tree, T is the number of trees, and ω is the regularisation term.

AdaBoost

AdaBoost works by adjusting all the weights without prior knowledge of weak learners [57]. The weakness of all the base learners is measured by the estimator’s error rate while training the models. Decision tree stumps are widely used with the AdaBoost algorithm to solve classification and regression problems. The objective function for the AdaBoost algorithm is defined using Eq 4.

L(H)=i=1Nexpyi*Hxi (4)

where H(xi) is the prediction of the classifier on the ith sample xi and yi is its corresponding true label in {-1, +1}and N denotes the number of training samples.

Gradient boosting

In this method, the weak learners are trained sequentially, and all estimators are added one by one by adapting the weights [58]. The gradient boosting algorithm focuses on predicting the residual errors of previous estimators and tries to minimise the difference between the predicted and actual values. The objective function for the gradient boosting algorithm is written using Eq 5.

L(θ)=minFi=1Nlyi,Fmathbfxi (5)

where F is the ensemble model, n is the number of training examples, yi is the true label of the ith sample, l denotes the loss function, and Fmathbf(xi) is the output of the ensemble model on example mathbf(xi).

4. Dataset collection and manipulation

We used the CKD data set (https://archive.ics.uci.edu/ml/datasets/chronic_kidney_disease), publicly available at the UCI machine learning repository, for the experiment. The dataset was collected from Apollo Hospitals, Managiri, India.

4.1 Dataset description

The dataset contains 400 instances and 25 attributes. The first 24 attributes are predicate/independent, and the last one is a dependent/target attribute. Among the attributes, 11 are numeric, and 14 are categorical. The attributes are described in Table 1. It represents the information about considered attributes, the description of attributes, their measurements, and the range values.

Table 1. Attributes information of the dataset.

Attribute Description Measurement Value range
Age (age) Participant’s age Years 2–90
Blood pressure (bp) Participant’s blood pressure mm/hg 50–180
Specific gravity (sg) Urine specific gravity of the participant Nominal 1.005–1.025
Albumin (al) Blood volume of the participant Nominal 0–5
Sugar (su) Participant’s sugar level in the blood Nominal 0–5
Red blood cells (rbc) Normality of red blood cells of the participant Categorical 0 or 1
Pus cell (pc) Normality of pus cells of the participant Categorical 0 or 1
Pus cell clumps (pcc) Presence of pus cell clumps in the participant’s urine Categorical 0 or 1
Bacteria (ba) Presence of bacteria in the participant’s urine Categorical 0 or 1
Blood glucose random (bgr) Blood sugar test of the participant mgs/dl 22–490
Blood urea (bu) Nitrogen level in the participant’s blood mgs/dl 1.50–391
Serum creatinine (sc) Creatinine level in the participant’s blood mgs/dl 0.40–76
Sodium (sod) Sodium level in the participant’s blood mEq/L 4.50–163
Potassium (pot) Potassium level in the participant’s blood mEq/L 2.50–47
Haemoglobin (hemo) Haemoglobin measure in the participant’s blood Gms 3.10–54
Packed cell volume (pcv) Measure and size of RBCs in the participant’s blood Numeric 9.00–54
White blood cell count (wc) WBCs count in the participant’s blood Cells/cumm 2200–26400
Red blood cell count (rc) RBCs count in the participant’s blood Millions/ cumm 2.10–8
Hypertension (htn) If the participant has hypertension Categorical 0 or 1
Diabetes mellitus (dm) If the participant has diabetes Categorical 0 or 1
Coronary artery disease (cad) If the participant has coronary artery disease Categorical 0 or 1
Appetite (appet) Participant’s desire or need for something to eat Categorical 0 or 1
Pedal edema (pe) If the participant has swelling in the ankles and feet Categorical 0 or 1
Anaemia (ane) Deficiency in RBCs of the participant Categorical 0 or 1
Class (outcome) If the participant has CKD Categorical 0 or 1

Table 2 describes the attribute information with their measures like count of records, mean, standard deviation (std), minimum (min) value, and maximum (max) value. For example, the blood pressure (bp) attribute has a count value of 400, mean 76.175, std 13.769, min 50, and max 180, respectively.

Table 2. Attributes information of the dataset.

Attribute Count Mean Std Min Max
Age (age) 400 51.585 17.308 2 90
Blood pressure (bp) 76.175 13.769 50 180
Specific gravity (sg) 1.017 0.005 1.005 1.025
Albumin (al) 1.057 1.343 0 5
Sugar (su) 0.450 1.084 0 5
Red blood cells (rbc) 0.727 0.445 0 1
Pus cell (pc) 0.773 0.420 0 1
Pus cell clumps (pcc) 0.105 0.307 0 1
Bacteria (ba) 0.055 0.228 0 1
Blood glucose random (bgr) 149.710 78.481 22 490
Blood urea (bu) 57.426 49.286 1.500 391
Serum creatinine (sc) 3.072 5.617 0.400 76
Sodium (sod) 136.790 10.039 4.500 163
Potassium (pot) 4.605 2.857 2.500 47
Haemoglobin (hemo) 12.332 2.926 3.100 17.80
Packed cell volume (pcv) 37.843 9.292 9 54
White blood cell count (wc) 8448 2951.563 2200 26400
Red blood cell count (rc) 4.473 1.009 2.100 8
Hypertension (htn) 0.368 0.483 0 1
Diabetes mellitus (dm) 0.343 0.475 0 1
Coronary artery disease (cad) 0.085 0.279 0 1
Appetite (appet) 0.795 0.404 0 1
Pedal edema (pe) 0.190 0.393 0 1
Anaemia (ane) 0.150 0.358 0 1
Class (outcome) 0.625 0.485 0 1

4.2 Data preprocessing

We performed some preprocessing on the considered CKD dataset to make the dataset most usable. The purpose was to transform the available raw data into a format easily understood by the ensemble learning algorithms. We conducted the following steps as data preprocessing:

  1. Identify and replace duplicate values.

  2. Identify and replace missing values.

  3. Detect and replace the outliers.

  4. Convert categorical variables to numerical values using one-hot encoding.

  5. Perform data transformation (-1 to 1) and scaling (0 to 1).

The results of the above steps are discussed below.

Class balancing

The training dataset should be balanced of positive and negative instances to achieve reasonable prediction. From Fig 2(A), it can be observed that the considered dataset was highly biased toward the positive class, i.e., “patients having CKD” over the negative class, “patients not having CKD.” To minimise this difference, we used SMOTE to balance the dataset. From Fig 2(B), it can be observed that the resultant dataset is fairly balanced.

Fig 2. Dataset balancing using SMOTE.

Fig 2

Exploratory data analysis

We used different data visualisation tools to visualise and analyse the distribution of the data samples. Fig 3 shows the normally distributed histograms that group all the attributes of the considered dataset within the range values. Here, the X- and Y-axes describe the input attributes, and their corresponding values, respectively. Fig 4 plots the probability density using the kernel density estimation (KDE) method. The X- and Y-axes denote each attribute’s parameter value and probability density function, respectively. Fig 5 depicts the boxplot of all the considered attributes of the dataset. It provides a good indication of how the dispersion of values is spread out. To handle the outliers in the dataset, the IQR method was used.

Fig 3. Histogram of the dataset attributes.

Fig 3

Fig 4. Density plot of the dataset attributes.

Fig 4

Fig 5. Boxplot of the dataset attributes.

Fig 5

Correlation coefficient analysis

To identify and plot the relationship among the dataset attributes, we used the correlation coefficient analysis (CCA) method. A strong association/relationship between the set of independent and dependent attributes indicates a good-quality dataset. Fig 6 presents the CCA of the dataset attributes used in the experiment. The relationship range lies between +1 to -1 along the X- and Y-axes.

Fig 6. Correlation coefficient analysis of the dependent and independent attributes in the dataset.

Fig 6

Data wrangling and cleaning

To clean the dataset, we identified the missing values using the isnull() method and then calculated the percentage of null values present in the dataset. We used the data imputation methods (mean, median, fill, and original) to replace the null values. The missing values were replaced using the column’s mean, median, and mode. We used the IRQ method to detect the outliers and replace them using the Z-score method. The Z-score method shifts the distribution of all the data samples and makes the mean 0. Using data cleaning methods, we further checked for duplicate, inconsistent, and corrupt values in the dataset and neutralised them wherever applicable.

Data standardisation and normalisation

We used the MinMaxScaler() for feature scaling. We scaled the data values using Eq 1 for standardisation and batch normalisation. Here, the data mean is set to 0 and the standard deviation to 6.

N(X)=i=1Nxixminxmaxxmin (6)

where N, X, xi, σ(x), xmin, and xmax denote the total sample in the data, ith attribute, the mean of the attributes, the sample variance of the attributes, the minimum value of the sample, and the maximum value of the sample, respectively.

5. Experiment, results, and discussion

In this section, we present the experimental details of this work and the obtained results by using the five boosting algorithms to predict CKD. We used 60% of the dataset to train the boosting algorithms and the rest 40% to test and validate their efficacy. The evaluations are extensively discussed in terms of accuracy, recall, precision, F1-score, micro-weighted, average-weighted, and AUC-ROC (area under curve-receiver operating characteristic) curve for each algorithm.

5.1 Hardware and software specifications

An HP Z60 workstation was used to carry out this research work. The hardware specification of the system is: Intel XEON 2.4 GHz CPU (12 core), 8 GB RAM, 1 T.B. hard disk, with Windows 10 pro-64-bit O.S. environment. As software requirements, we used the GUI-based Anaconda Navigator, the web-based computing platform Jupyter notebook, and Python as the programming language.

5.2 Feature importance

The feature importance is used to assess the contribution of an independent/predicate attribute that leads to CKD. Generally, not all attributes contribute to disease prediction. For instance, after running all five boosting algorithms on the original dataset, we found that the attributes–‘ane’, ‘appet’, ‘ba’, ‘cad’, ‘pc’, ‘pcc’, ‘pe’, ‘su’, and ‘wc’ have no role in CKD prediction. Hence, we eliminated these attributes from the dataset and kept only those that contributed at least for one algorithm, as shown in Fig 7.

Fig 7. Contributing features in CKD prediction for all boosting algorithms.

Fig 7

We used the forward selection, a wrapper method, to calculate the feature importance [59]. A higher F-score of a feature indicates more importance of an attribute. For example, in Fig 7, it can be seen that the haemoglobin (hemo) attribute has the highest contribution in the CKD prediction for all the algorithms.

5.3 Hyperparameter tuning

We used the grid search method for hypermeter tuning to achieve optimality in the proposed model’s performance. By specifying a grid or a specified set of values for each hyperparameter, grid search enables methodically examining various combinations of hyperparameters. This ensures that all the options are tried to find the optimal values of the hyperparameters. The deterministic nature of grid search ensures consistency, i.e., it always yields the same outcomes when the same hyperparameters and data are used. This characteristic facilitates transparent testing and evaluation by making results simple to replicate and compare. One of the major advantages of grid search is that it is fairly straightforward to implement. Also, most of the machine learning frameworks and libraries provide built-in functions or modules for grid search. The best values of the hyperparameters found for each algorithm are shown in Table 3. The listed values for each parameter for the respective algorithm were found to be the best performers in our experiment.

Table 3. The optimal hyperparameters of boosting algorithms.

Boosting algorithm Hyperparameters
XGBoost XGBClassifier (learning_rate = 0.1, n_estimators = 1000, max_depth = 5, min_child_weight = 6, ’reg_alpha’: 60.0, subsample = 0.6, colsample_bytree = 0.8, ’gamma’: 4.20).
CatBoost CatBoostClassifier (random_state = 0, learning_rate = [0.1, 0.05], n_estimators = 100, max_depth = [1,3,5], leaf_reg’, 2.0, 8, 16, min_child_samples = 2, 4, 6,
LightGBM LightGBM (boosting_type = ’lgbm’, random_state = 45, learning_rate = 0.1, n_estimators = 1000, max_depth = 2, min_child_samples = 250, silent = True, n_jobs = 6).
AdaBoost GridSearchCV (random_state = 45, learning_rate = [0.01, 0.05], n_estimators = 200, algorithm = ’SAMME.R’, n_jobs = n_jobs).
Gradient boosting GridSearchCV (random_state = 45, learning_rate = [0.1, 2, 5], estimators = GradientBoostingClassifier(), max_depth = 4, weight = 6, verbose = 1).

5.4 Cross-validation scheme

Cross-validation is conducted to provide an unbiased evaluation of the prediction model. We performed the k-fold cross-validation to validate the performance of the proposed model on the training dataset. Here, we kept the value of k as 6. Based on the validation bias, the hyperparameters used in the experiment were tuned.

5.5 Performance evaluation

In this section, the performance of the proposed prediction model for the considered boosting algorithms is discussed in terms of different performance metrics.

5.5.1 Classification accuracy

The classification performances of the algorithms are evaluated using a confusion matrix. The confusion matrices of all five boosting algorithms applied on the test dataset are shown in Fig 8. The left upper and the right lower boxes denote the correct predictions for the patients having (true positive) and not having (true negative) CKD, respectively. The right upper box and the left lower box indicate the number of incorrect predictions for patients having (false positive) and not having (false negative) CKD, respectively. The training and testing accuracies of all the boosting algorithms are shown in Fig 9. As per our experiment, on the test dataset, AdaBoost outperformed other algorithms by attaining the maximum accuracy rate for the training set of 100% and the test set of 98.47%, followed by LightGBM, gradient boost, XGBoost, and CatBoost at 99.73%, 99.21%, 97.23%, and 96.97%, respectively on the training set, and 97.96%, 97.46%, 95.93%, and 96.44%, respectively on the test set.

Fig 8. Confusion matrices of the prediction performance on the test set for all the five boosting algorithms.

Fig 8

Fig 9. Training and testing accuracy statistics of all the boosting algorithms.

Fig 9

5.5.2 Other measurements

In addition to accuracy, we calculated the precision, recall, F1-score, and support of the five boosting algorithms on the test set, as shown in Figs 1013, respectively. In addition, the macro and weighted average were measured for both classes (0: no CKD, 1: CKD). As shown in those figures, AdaBoost produced the best precision in identifying the presence of CKD, while all algorithms identified the non-infection of CKD with equal precision. AdaBoost has a better recall and F1-score in confirming the absence of CKD. Regarding the case of support, i.e., the occurrence of class, AdaBoost performs slightly better than the other algorithms.

Fig 10. Precision comparison of five boosting algorithms on test set.

Fig 10

Fig 13. Support comparison of five boosting algorithms on test set.

Fig 13

Fig 11. Recall comparison of five boosting algorithms on test set.

Fig 11

Fig 12. F1-score comparison of five boosting algorithms on test set.

Fig 12

5.5.3 AUC-ROC curve

The AUC-ROC curve was used to show the prediction ability of the boosting algorithms at different thresholds. It represents a false-positive rate (FPR) vs. a true-positive rate (TPR) along the x-axis and y-axis. A larger AUC-ROC area suggests the model’s ability to distinguish between 0’s and 1’s, leading to a better prediction. Also, an AUC value closer to 1 denotes a good separability measure, while in the case of an AUC of below 0.5, the model becomes ineffective in separating the classes, denoting the bad measure of disassociation. The AUC-ROC for the experiment is shown in Fig 14. It can be observed that AdaBoost performs best while XGBoost performs worst.

Fig 14. AUC-ROC curves for the experimented boosting algorithms.

Fig 14

6. Comparative analysis

Table 4 presents a comparative analysis of the five boosting algorithms applied on the test dataset in terms of accuracy, misclassification rate, and runtimes. It can be observed that AdaBoost has the highest accuracy and least misclassification rate, but it has a slightly higher runtime than LightGBM and XGBoost.

Table 4. Comparative analysis of the considered algorithms performed on the test set.

Algorithm Accuracy (%) Misclassification rate (%) Runtime (seconds)
XGB 95.93 4.07 1.215
CatBoost 96.44 3.56 2.009
LGBM 97.96 2.04 1.005
ADB 98.47 1.53 1.970
GB 97.46 2.54 2.752

Since, in our experiment, we found AdaBoost to have the best overall performance in predicting CKD, we compared it with a few related research works in terms of accuracy, as shown in Table 5. The justification for achieving higher accuracy can be credited to the adopted procedures like data imputation for handling missing values, detection and replacing outliers, and effective data standardisation and normalisation.

Table 5. Comparison of the proposed work with existing similar works.

Research work Ensemble techniques adopted Dataset used Highest accuracy Precision Recall AUC/ ROC
Jongbo et al. [1] Individual + bagging ensemble approach + random subspace ensemble (naive Bayes, kNN, and decision tree) Chronic Kidney Dataset collected from UCI machine learning repository 98.30% with decision tree using random subsample ensemble - 98.50% 100%
Kumar et al. [42] SVM, C4.5 decision tree, PSO-MLP, decision tree, and artificial bee colony C4.5 92.76% with artificial bee colony 4.5 0.57% 0.42% -
Saurabh Pal [43] Logistic regression, decision tree, SVM, and bagging method 95.92% with decision tree 99% 98% -
Hasan and Hasan [44] AdaBoost, bootstrap aggregating, extra trees, gradient boosting, and random forest 99% with AdaBoost 98% 100% 99%
Wibawa et al. [45] AdaBoost based on KNN 98.01% with AdaBoost 97.86% 97.83% -
Ebiaredoh-Mienye et al. [46] Logistic regression, decision tree, XGBoost, random forest, SVM, and CS AdaBoost 99.80% with C.S. AdaBoost 97.50% 100% 98%
Emon et al. [47] Logistic regression, naive Bayes, multilayer perceptron, stochastic gradient descent, adaptive boosting, bagging, decision tree, and random forest 99% with Random forest 98.50% 98.50% 98%
Ramaswamyreddy et al. [48] Tree bag, AdaBoost, gradient boosting, and random forest 99% with gradient boosting - - -
Our method XGB, CatBoost, LGBM, ADB, and gradient boosting 98.47% with ADB 98.50% 98.50% 98.60%

7. Conclusion, limitations, and future directions

Diagnosis and prevention of chronic kidney disease have become challenging for healthcare professionals and other concerned authorities. It can be mitigated to some extent if it can be pre-diagnosed in well advance. In this paper, we attempted to predict CKD using an ensemble learning approach. Specifically, we used five boosting algorithms: XGBoost, CatBoost, LightGBM, AdaBoost, and gradient boosting. We employed different preprocessing techniques like the imputation method for handling missing values and min-max scalar and Z-score for data standardisation and normalisation. In addition, hyperparameter techniques like grid search were used to find the optimal parameter values. Furthermore, feature selection was carried out for each algorithm. AdaBoost emerged as the overall best performer in accuracy (99.17%), precision, recall, f1-score, and support in the experiment. AdaBoost also attained better results for AUC-ROC and misclassification rate. Comparing our proposed model with similar works, we found that our method outperformed others.

Though the proposed model performed relatively well, it has some obvious limitations. The size of the considered dataset is small, which may limit the prediction model’s performance in generic situations. It is observed that most of the features are having least contribution towards CKD. A more balanced dataset would lead to a better prediction model.

As an extension of this work, other ensemble learning techniques, like bagging, stacking, etc., can be explored to improve the results. Additionally, deep learning techniques can also be experimented with the exercised dataset. To validate the effectivity of the proposed model, additional and larger datasets are needed in future. Our proposed model can be applied to other disease datasets (e.g., diabetes) with common features. We expect more powerful disease prediction models to be developed and implemented in medical diagnosis and treatment.

Data Availability

All relevant data can be found at https://archive.ics.uci.edu/ml/datasets/chronic_kidney_disease.

Funding Statement

Zhongming Zhao was partially supported by his startup fund from The University of Texas Health Science Center at Houston, Houston, Texas, USA.

References

  • 1.Jongbo O. A., Adetunmbi A. O., Ogunrinde R. B. and Badeji-Ajisafe B., “Development of an ensemble approach to chronic kidney disease diagnosis,” Scientific African, vol. 8 (Article Id e00456), 2020. [Google Scholar]
  • 2.Rhee C. M. and Kovesdy C. P., “Epidemiology: spotlight on CKD deaths-increasing mortality worldwide,” Nature Reviews Nephrology, vol. 11, pp. 199–200, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Ravani P., Quinn R., Fiocco M., Liu P., Al-Wahsh H., Lam N., Hemmelgarn B. R., Manns B. J., James M. T., Joanette Y. and Tonelli M., “Association of age with risk of kidney failure in adults with stage IV chronic kidney disease in Canada,” JAMA Network Open, vol. 3, no. 9, p. e2017150, 2020. doi: 10.1001/jamanetworkopen.2020.17150 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Liu P., Quinn R. R., Lam N. N., Al-Wahsh H., Sood M. M., Tangri N., Tonelli M. and Ravani P., “Progression and regression of chronic kidney disease by age among adults in a population-based cohort in Alberta, Canada,” JAMA Network Open, vol. 4, no. 6, p. e2112828, 2021. doi: 10.1001/jamanetworkopen.2021.12828 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.“World Population Ageing 2019,” United Nations, Department of Economic and Social Affairs, Population Division, 31 December 2019. [Online]. Available: https://www.un.org/en/development/desa/population/publications/pdf/ageing/WorldPopulationAgeing2019-Report.pdf. [Accessed 24 September 2022].
  • 6.Bello A. K., Levin A., Tonelli M., Okpechi I. G., Feehally J., Harris D., Jindal K., Salako B. L., Rateb A., Osman M. A., Qarni B., Saad S., Lunney M., Wiebe N., Ye F. and Johnson D. W., “Assessment of global kidney health care status,” JAMA, vol. 317, no. 18, pp. 1864–1881, 2017. doi: 10.1001/jama.2017.4046 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Centers for Disease Control and Prevention, “National Chronic Kidney Disease Fact Sheet, 2017,” 2017.
  • 8.Centers for Disease control and Prevention, “Chronic Kidney Disease in the United States, 2021,” 2021.
  • 9.Levey A. S. and Coresh J., “Chronic kidney disease,” Lancet, vol. 379, no. 9811, pp. 165–180, 2012. doi: 10.1016/S0140-6736(11)60178-5 [DOI] [PubMed] [Google Scholar]
  • 10.Kovesdy C. P., “Epidemiology of chronic kidney disease: an update 2022,” Kidney International Supplements, vol. 12, no. 1, pp. 7–11, 2022. doi: 10.1016/j.kisu.2021.11.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.KDIGO, “Kidney Disease: Improving Global Outcomes (KDIGO) CKD Work Group KDIGO 2012 clinical practice guideline for the evaluation and management of chronic kidney disease,” Kidney International Supplements, vol. 3, no. 1, 2013.
  • 12.Yuan Q., Zhang H., Deng T., Tang S., Yuan X., Tang W., Xie Y., Ge H., Wang X., Zhou Q. and Xiao X., “Role of Artificial Intelligence in Kidney Disease,” International Journal of Medical Sciences, vol. 17, no. 7, pp. 970–984, 2020. doi: 10.7150/ijms.42078 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Chan L., Vaid A. and Nadkarni G. N., “Applications of machine learning methods in kidney disease: Hope or Hype?,” Current Opinion in Nephrology and Hypertension, vol. 29, no. 3, pp. 319–326, 2020. doi: 10.1097/MNH.0000000000000604 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.P. K. D. Pramanik, M. Mukhopadhyay and S. Pal, “Big data classification: applications and challenges,” in Artificial Intelligence and IoT: Smart Convergence for Eco Friendly Topography. Studies in Big Data, vol. 85, M. K. Geetha, J. A. Nehru and B. Sivaraman, Eds., Singapore, Springer, 2021, pp. 53–84.
  • 15.Ganie S. M., Malik M. B. and Arif T., “Performance analysis and prediction of type 2 diabetes mellitus based on lifestyle data using machine learning approaches,” Journal of Diabetes & Metabolic Disorders, vol. 21, no. 1, pp. 339–352, 2022. doi: 10.1007/s40200-022-00981-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Hazra A., Mandal S. K., Gupta A., Mukherjee A. and Mukherjee A., “Heart disease diagnosis and prediction using machine learning and data mining techniques: a review,” Advances in Computational Sciences and Technology, vol. 10, no. 7, pp. 2137–2159, 2017. [Google Scholar]
  • 17.Nissa N., Jamwal S. and Mohammad S., “Early detection of cardiovascular disease using machine learning techniques an experimental study,” International Journal of Recent Technology and Engineering (IJRTE), vol. 9, no. 3, pp. 635–641, 2020. [Google Scholar]
  • 18.Senan E. M., Al-Adhaileh M. H., Alsaade F. W., Aldhyani T. H. H., Alqarni A. A., Alsharif N., Uddin M. I., Alahmadi A. H., Jadhav M. E. and Alzahrani M. Y., “Diagnosis of chronic kidney disease using effective classification algorithms and recursive feature elimination techniques,” Journal of Healthcare Engineering, vol. 2021, 2021. doi: 10.1155/2021/1004767 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Abdel-Fattah M. A., Othman N. A. and Goher N., “Predicting chronic kidney disease using hybrid machine learning based on Apache Spark,” Computational Intelligence and Neuroscience, vol. 2022, p. 9898831, 2022. doi: 10.1155/2022/9898831 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Bai Q., Su C., Tang W. and Li Y., “Machine learning to predict end stage kidney disease in chronic kidney disease,” Scientific Reports, vol. 12 (Article number: 8377), 2022. doi: 10.1038/s41598-022-12316-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Ilyas H., Ali S., Ponum M., Hasan O., Mahmood M. T., Iftikhar M. and Malik M. H., “Chronic kidney disease diagnosis using decision tree algorithms,” BMC Nephrology, vol. 22 (Article number: 273), 2021. doi: 10.1186/s12882-021-02474-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Yashfi S. Y., Islam M. A., Pritilata N. Sakib T. Islam M. Shahbaaz and Pantho S. S., “Risk prediction of chronic kidney disease using machine learning algorithms,” in 11th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kharagpur, India, 2020. [Google Scholar]
  • 23.Onyema E. M., Almuzaini K. K., Onu F. U., Verma D., Gregory U. S., Puttaramaiah M. and Afriyie R. K., “Prospects and Challenges of Using Machine Learning for Academic Forecasting,” Computational Intelligence and Neuroscience, vol. 2022 (Article ID 5624475), 2022. doi: 10.1155/2022/5624475 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Ghassemi M., Naumann T., Schulam P., Beam A. L., Chen I. Y. and Ranganath R., “A Review of Challenges and Opportunities in Machine Learning for Health,” AMIA Joint Summits on Translational Science Proceedings, vol. 2020, pp. 191–200, 2020. doi: 10.1001/jama.2017.18391 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Ganie S. M. and Malik M. B., “An ensemble machine learning approach for predicting type-II diabetes mellitus based on lifestyle indicators,” Healthcare Analytics, vol. 22 (Article number: 100092), 2022. [Google Scholar]
  • 26.Naveen R. K. Sharma and Nair A. R., “Efficient breast cancer prediction using ensemble machine learning models,” in 4th International Conference on Recent Trends on Electronics, Information, Communication & Technology (RTEICT), Bangalore, India, 2019. [Google Scholar]
  • 27.Oswald G. J. Sathwika and Bhattacharya A., “Prediction of cardiovascular disease (CVD) using ensemble learning algorithms,” in 5th Joint International Conference on Data Science & Management of Data (9th ACM IKDD CODS and 27th COMAD), Bangalore, India, 2022. [Google Scholar]
  • 28.Mienye I. D., Sun Y. and Wang Z., “An improved ensemble learning approach for the prediction of heart disease risk,” Informatics in Medicine Unlocked, vol. 20 (100402), 2020. [Google Scholar]
  • 29.Shanbhag G. A., Prabhu K. A., Subba Reddy N. V. and Rao B. A., “Prediction of lung cancer using ensemble classifiers,” Journal of Physics: Conference Series, vol. 2161 (012007), 2022. [Google Scholar]
  • 30.Verma A. K., Pal S. and Tiwari B. B., “Skin disease prediction using ensemble methods and a new hybrid feature selection technique,” Iran Journal of Computer Science, vol. 3, pp. 207–216, 2020. [Google Scholar]
  • 31.Jia P., Hu R. and Zhao Z., “Benchmark of embedding-based methods for accurate and transferable prediction of drug response,” Briefings in Bioinformatics, p. bbad098, 2023. doi: 10.1093/bib/bbad098 [DOI] [PubMed] [Google Scholar]
  • 32.Ganie S. M. and Malik M. B., “Comparative analysis of various supervised machine learning algorithms for the early prediction of type-II diabetes mellitus,” International Journal of Medical Engineering and Informatics, vol. 14, no. 6, pp. 473–483, 2022. [Google Scholar]
  • 33.Shaikh F. J. and Rao D. S., “Prediction of cancer disease using machine learning approach,” Materialstoday: Proceedings, vol. 50 (Part 1), pp. 40–47, 2022. [Google Scholar]
  • 34.Latha C. B. C. and Jeeva S. C., “Improving the accuracy of prediction of heart disease risk based on ensemble classification techniques,” Informatics in Medicine Unlocked, vol. 16 (100203), 2019. [Google Scholar]
  • 35.Ali F., El-Sappagh S., Islam S. R., Kwak D., Ali A., Imran M. and Kwak K.-S, “A smart healthcare monitoring system for heart disease prediction based on ensemble deep learning and feature fusion,” Information Fusion, vol. 63, pp. 208–222, 2020. [Google Scholar]
  • 36.Senthilkumar B., Zodinpuii D., Pachuau L., Chenkual S., Zohmingthanga J., Kumar N. c. S and Hmingliana L., “Ensemble modelling for early breast cancer prediction from diet and lifestyle,” IFAC-PapersOnLine, vol. 55, no. 1, pp. 429–435, 2022. [Google Scholar]
  • 37.Verma A. K., Pal S. and Kumar S., “Comparison of skin disease prediction by feature selection using ensemble data mining techniques,” Informatics in Medicine Unlocked, vol. 16 (100202), 2019. [Google Scholar]
  • 38.Yadav D. C. and Pal S., “Prediction of thyroid disease using decision tree ensemble method,” Human-Intelligent Systems Integration, vol. 2, pp. 89–95, 2020. [Google Scholar]
  • 39.Hakim M. A., Jahan N., Zerin Z. A. and Farha A. B., “Performance evaluation and comparison of ensemble based bagging and boosting machine learning methods for automated early prediction of myocardial infarction,” in 12th International Conference on Computing Communication and Networking Technologies (ICCCNT), Kharagpur, India, 2021. [Google Scholar]
  • 40.El-Sappagh S., Ali F., Abuhmed T., Singh J. and Alonso J. M., “Automatic detection of Alzheimer’s disease progression: An efficient information fusion approach with heterogeneous ensemble classifiers,” Neurocomputing, vol. 512, pp. 203–224, 2022. [Google Scholar]
  • 41.Lim D. K. E., Boyd J. H., Thomas E., Chakera A., Tippaya S., Irish A., Manuel J., Betts K. and Robinson S., “Prediction models used in the progression of chronic kidney disease: A scoping review,” PLoS ONE, vol. 17, no. 7, p. e0271619, 2022. doi: 10.1371/journal.pone.0271619 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Komal Kumar N., Tulasi R. L. and Vigneswari D., “An ensemble multi-model technique for predicting chronic kidney disease,” International Journal of Electrical and Computer Engineering (IJECE), vol. 9, no. 2, pp. 1321–1326, 2019. [Google Scholar]
  • 43.Pal S., “Chronic kidney disease prediction using machine learning techniques,” Biomedical Materials & Devices, 2022. [Google Scholar]
  • 44.Hasan K. M. Z. and Hasan M. Z., “Performance evaluation of ensemble-based machine learning techniques for prediction of chronic kidney disease,” in Emerging Research in Computing, Information, Communication and Applications. Advances in Intelligent Systems and Computing, vol. 882, Shetty N., Patnaik L., Nagaraj H., Hamsavath P. and Nalini N., Eds., Singapore, Springer, 2019, pp. 415–426. [Google Scholar]
  • 45.Wibawa M. S., Maysanjaya I. M. D. and Putra I. M. A. W., “Boosted classifier and features selection for enhancing chronic kidney disease diagnose,” in 5th International Conference on Cyber and IT Service Management (CITSM), Denpasar, Indonesia, 2017. [Google Scholar]
  • 46.Ebiaredoh-Mienye S. A., Swart T. G., Esenogho E. and Mienye I. D., “A machine learning method with filter-based feature selection for improved prediction of chronic kidney disease,” Bioengineering, vol. 9, no. 8, p. 350, 2022. doi: 10.3390/bioengineering9080350 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Emon M. U., Imran A. M., Islam R., Keya M. S., Zannat R. and Ohidujjaman, “Performance analysis of chronic kidney disease through machine learning approaches,” in 6th International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 2021. [Google Scholar]
  • 48.Ramaswamyreddy A., Shiva Prasad S., Rangarao K. V. and Saranya A., “Efficient datamining model for prediction of chronic kidney disease using wrapper methods,” International Journal of Informatics and Communication Technology, vol. 8, no. 2, pp. 63–70, 2019. [Google Scholar]
  • 49.G. Brown, “Ensemble learning,” in Encyclopedia of Machine Learning, C. Sammut and G. I. Webb, Eds., Boston, MA, Springer, 2011, pp. 312–320.
  • 50.Sagi O. and Rokach L., “Ensemble learning: a survey,” WIREs Data Mining and Knowledge Discovery, vol. 8, no. 4, p. e1249, 2018. [Google Scholar]
  • 51.Zhang C. and Ma Y., Eds., Ensemble machine learning: methods and applications, New York: Springer, 2012. [Google Scholar]
  • 52.Ferreira A. J. and Figueiredo M. A. T., “Boosting algorithms: a review of methods, theory, and applications,” in Ensemble Machine Learning, Zhang C. and Ma Y., Eds., Boston, MA, Springer, 2012, pp. 35–85. [Google Scholar]
  • 53.Tanha J., Abdi Y., Samadi N., Razzaghi N. and Asadpour M., “Boosting methods for multi-class imbalanced data classification: an experimental review,” Journal of Big Data, vol. 7, no. 1, 2020. [Google Scholar]
  • 54.Santhanam R., Uzir N., Raman S. and Banerjee S., “Experimenting XGBoost algorithm for prediction and classification of different datasets,” International Journal of Control Theory and Applications, vol. 9, no. 40, pp. 651–662, 2016. [Google Scholar]
  • 55.Hancock J. T. and Khoshgoftaar T. M., “CatBoost for big data: an interdisciplinary review,” Journal of Big Data, vol. 7, no. 1 (Article No. 94), 2020. doi: 10.1186/s40537-020-00369-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Machado M. R., Karray S. and de Sousa I. T., “LightGBM: an effective decision tree gradient boosting method to predict customer loyalty in the finance industry,” in International Conference on Computer Science & Education (ICCSE), Toronto, Canada, 2019. [Google Scholar]
  • 57.Sevinc E., “An empowered AdaBoost algorithm implementation: A COVID-19 dataset study,” Computers & Industrial Engineering, vol. 165 (Article No. 107912), 2022. doi: 10.1016/j.cie.2021.107912 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Aziz N., Akhir E. A. P., Aziz I. A., Jaafar J., Hasan M. H. and Abas A. N. C., “A study on gradient boosting algorithms for development of AI monitoring and prediction systems,” in International Conference on Computational Intelligence (ICCI), Malaysia, 2020. [Google Scholar]
  • 59.Kohavi R. and John G. H., “Wrappers for feature subset selection,” Artificial Intelligence, vol. 97, no. 1–2, pp. 273–324, 1997. [Google Scholar]

Decision Letter 0

Muhammad Fazal Ijaz

24 Feb 2023

PONE-D-23-02469Chronic Kidney Disease Prediction Using Boosting Techniques based on Clinical ParametersPLOS ONE

Dear Dr. Ganie,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Apr 10 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Muhammad Fazal Ijaz

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1.  Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf  and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Partly

Reviewer #3: No

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: I Don't Know

Reviewer #3: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: No

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The overall impression of the technical contribution of the current study is reasonable. However, the Authors may consider making tremendous amendments to the manuscript for better comprehensibility of the study.

1. The abstract must be re-written, focusing on the technical aspects of the proposed model, the main experimental results, and the metrics used in the evaluation. Briefly discuss how the proposed model is superior.

2. Please make sure the abbreviations are properly used. For example Area Under Curve-Receiving Operator Characteristic (AUC-ROC).

3. Additionally, method names should not be capitalized. Moreover, it is not the best practice to employ abbreviations in the abstract, they should be used when the term is introduced for the first time.

4. This paper is not a survey paper to include the references as "[30] [31] [32] [33] [34]." Please provide the significance of each of those studies and add a citation.

5. Literature must be tremendously improvised for better idea on the field of study and state of art models, authors may consider some of the relevant studies like https://doi.org/10.1371/journal.pone.0271619

6. Boosting algorithms must be adequately explain, for better idea refer and include https://doi.org/10.3390/healthcare10010085

7. How are the feature weights calculated, what was the approach or the technique that is being used.

8. Manuscript is having too many sub-sections, please minimize for better readability of the study.

9. The section Data Standardization and Normalisation must be discussed as data pre-processing phase, that has to be in background section of the manuscript.

10. what are the cases that are assumed as TP, TN, FP and FN, please explain them clearly, for better idea refer https://doi.org/10.3390/s21082852

11. Authors may present the loss functions for better comprehensibility of each of the models used in the proposed model. For better idea refer https://doi.org/10.3390/s21165386

12. Majority of the figures lack the clarity, they quality is fair but they must be explained in the text and the figures must be cited.

13. More comparative analysis with state-of-art models is desired.

14. By considering the current form of the conclusion section, it is hard to understand by PLos One Journal readers. It should be extended with new sentences about the necessity and contributions of the study by considering the authors' opinions about the experimental results derived from some other well-known objective evaluation values if it is possible.

15. English proofreading is strongly recommended for a better understanding of the study, and the quality of the figures must be tremendously improved.

Reviewer #2: This paper presents discussion on image systems for medical systems data analytics.

1. What are main aspects of novelty and advances in the described ideas?

2. What are limitations of presented approaches? How the models work in different scenarios of operation? What are weak points of presented ideas?

3. Related ideas: Deep neural network correlation learning mechanism for CT brain tumor detection, BiLSTM deep neural network model for imbalanced medical data of IoT systems, Deep learning for neurodegenerative disorder (2016 to 2022): A systematic review, .

4. Compare model to other in different operation and in different positioning of the input data.

5. What are future trends in the development of this type systems? How the development would work in different configurations? What kind of transfer and network configurations are necessary for your model?

6. How to set optimal coefficients for these models? Did you test other configurations? How were these selected?

7. Did you test the option to transform the knowledge before processing?

8. There are no comparisons to other models so we are not able to see advances of your processing.

9. Your fig 1 is not much informative since there are no details on your model thus we are not able to repeat your experiment.

10. How do you understand T in your model? Is this a time of processing or number of iterations?

Reviewer #3: In this paper, authors presented boosting techniques for chronic kidney disease prediction. However, there are some limitations that must be addressed as follows.

1. There are some typos and grammatical errors in Abstract. In addition, the abstract is not attractive. Some sentences in abstract should be modified to make it more attractive for readers.

2. In Introduction section, it is difficult to understand the novelty of the presented research work. In addition, some references are missing.

3. In related work, the existing works about patient disease prediction should be discussed: ‘A smart healthcare monitoring system for heart disease prediction based on ensemble deep learning and feature fusion’, ‘Alzheimer’s disease progression detection model based on an early fusion of cost-effective multimodal data’, Automatic detection of Alzheimer’s disease progression: An efficient information fusion approach with heterogeneous ensemble classifiers’, and ‘An intelligent healthcare monitoring framework using wearable sensors and social networking data.’

4. The authors should properly select and check the subsection title. There are so many typos and (see section 3 research methodolog.

5. The number given to each section is not correct.

6. Where are the other preprocessing steps? How is the data preprocessed?

7. What about feature selection?

8. The results are not properly discussed.

9. Captions of the Figures not self-explanatory. The caption of figures should be self-explanatory, and clearly explaining the figure. Extend the description of the mentioned figures to make them self-explanatory.

10. Equation 2 should be more clearly discussed.

11. In conclusion section, the future work should be more deeply discussed.

12. The whole manuscript should be thoroughly revised in order to improve its English.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Dec 1;18(12):e0295234. doi: 10.1371/journal.pone.0295234.r002

Author response to Decision Letter 0


8 Mar 2023

Response to Reviewers’ comments

Reviewer 1

1. The abstract must be re-written, focusing on the technical aspects of the proposed model, the main experimental results, and the metrics used in the evaluation. Briefly discuss how the proposed model is superior

Response: We thank the reviewer for the suggestion. The abstract is written including the suggested points.

2. Please make sure the abbreviations are properly used. For example, Area Under Curve-Receiving Operator Characteristic (AUC-ROC).

Response: We apologies for the silly mistake. It is duly corrected.

3. Additionally, method names should not be capitalized. Moreover, it is not the best practice to employ abbreviations in the abstract, they should be used when the term is introduced for the first time.

Response: We acknowledge the reviewer’s concern regarding capitalization. However, in the literature the normal convention to write the following terms are: XGBoost, CatBoost, LightGBM, AdaBoost.

Furthermore, these terms are more popularly known by their abbreviated forms only. Writing the full forms for each of them would increase the word count of the abstract. Nevertheless, as suggested by the reviewer, the full forms of these terms are given in the text when they are used for first time.

4. This paper is not a survey paper to include the references as "[30] [31] [32] [33] [34]." Please provide the significance of each of those studies and add a citation.

Response: We thank the reviewer for the suggestion. The goal of the sentence is to establish the importance of ensemble learning in disease diagnosis and treatment. As suggested, we re- written the sentence by including sample research works that addressed different diseases using ensemble learning.

5. Literature must be tremendously improvised for better idea on the field of study and state of art models, authors may consider some of the relevant studies like https://doi.org/10.1371/journal.pone.0271619

Response: We thank the reviewer for the suggestion. We included the suggested paper in the related work. We also improved the literature survey following the paper. A couple of recent papers are included in the related work.

6. Boosting algorithms must be adequately explain, for better idea refer and include https://doi.org/10.3390/healthcare10010085

Response: We thank the reviewer for the suggestion. The boosting algorithms are elaborated with suitable mathematical equations.

7. How are the feature weights calculated, what was the approach or the technique that is being used.

Response: To calculate the feature importance, we used the wrapper method. This is included in the manuscript with suitable reference.

8. Manuscript is having too many sub-sections, please minimize for better readability of the study.

Response: We thank the reviewer for bringing our notice into this. There was a problem with the subsection 3.1 and Section 4 which should be included under Section 3. It is corrected duly. We structured the manuscript into different subsections which allowed us to express the non-overlapping contents precisely and exclusively. This would help the readers to focus and understand the fragments of the experiment in a granular way.

9. The section Data Standardization and Normalisation must be discussed as data pre-processing phase, that has to be in background section of the manuscript.

Response: As suggested, the section Data Standardization and Normalisation is included in the Data Pre-processing section.

10. What are the cases that are assumed as TP, TN, FP and FN, please explain them clearly, for better idea refer https://doi.org/10.3390/s21082852

Response: We thank the reviewer for the suggestion. Each prediction cases are stated with respect to Figure 7.

11. Authors may present the loss functions for better comprehensibility of each of the models used in the proposed model. For better idea refer https://doi.org/10.3390/s21165386

Response: The misclassification rate of incorrect prediction is given in Table 4.

12. Majority of the figures lack the clarity, they quality is fair but they must be explained in the text and the figures must be cited.

Response: The figures are described and cited in the text.

13. More comparative analysis with state-of-art models is desired.

Response: As suggested by the reviewer, our work has been compared with a couple of papers in addition to the existing 6 papers, as shown in Table 5.

14. By considering the current form of the conclusion section, it is hard to understand by PLos One Journal readers. It should be extended with new sentences about the necessity and contributions of the study by considering the authors' opinions about the experimental results derived from some other well-known objective evaluation values if it is possible

Response: The necessity and contribution of this research work is included in the Conclusion section.

15. English proofreading is strongly recommended for a better understanding of the study, and the quality of the figures must be tremendously improved.

Response: The manuscript is proofread for possible grammatical and writing mistakes. Regarding the figures, most of them are program generated and large in size. Accommodating in a smaller scale makes a couple of figures look unclear. However, they are perfectly readable by zooming.

Reviewer 2

1. What are main aspects of novelty and advances in the described ideas?

Response: In this paper, we proposed a novel CKD prediction model using ensemble learning. We used five different boosting algorithms to check the prediction performance of the model. Along with achieving better results (e.g., accuracy, precision, recall, etc.) and runtime we also assessed the contributions of all the attributes in the dataset that cause CKD.

2. What are limitations of presented approaches? How the models work in different scenarios of operation? What are weak points of presented ideas?

Response: We thank the reviewer for the suggestion. The limitations of the work in included in the Conclusion section.

3. Related ideas: Deep neural network correlation learning mechanism for CT brain tumor detection, BiLSTM deep neural network model for imbalanced medical data of IoT systems, Deep learning for neurodegenerative disorder (2016 to 2022): A systematic review.

Response: We thank the reviewer for the valuable suggestion. We’ve already started working on the deep learning based models for different disease predictions.

4. Compare model to other in different operation and in different positioning of the input data.

Response: We tried with different combinations of the data preprocessing, feature selection and hyperparameter tuning and came out with the best performing model. The model was tried with five boosting algorithms. Finally, the performance of the best working combination of the proposed model with AdaBoost is compared with a number of state-of-the-art findings.

5. What are future trends in the development of this type systems? How the development would work in different configurations? What kind of transfer and network configurations are necessary for your model?

Response: In future, more sophisticated disease prediction models will be prevalent in the medical diagnosis and treatment. The proposed model can be used for other healthcare datasets that share the commonality of features. The configuration setup will depend on the particular application requirement and the properties of the available dataset. The future works of this work is mentioned in the Conclusion section.

6. How to set optimal coefficients for these models? Did you test other configurations? How were these selected?

Response: We tested the performance of the prediction model for combination of different coefficient values for all the tunable parameters. Among them, the optimal value sets were selected, as shown in Table 3.

7. Did you test the option to transform the knowledge before processing?

Response: We carried out the background work to set up the base prediction model and tested it on different ensemble algorithms. Among them, in the given setup, AdaBoost performed best.

8. There are no comparisons to other models so we are not able to see advances of your processing.

Response: The proposed model is compared with several similar published works, as shown in Table 5. It can be observed that our model outperforms the other compared works.

9. Your fig 1 is not much informative since there are no details on your model thus we are not able to repeat your experiment.

Response: Fig. 1 shows only the overall flow of the paper. The details of each step are elaborately discussed in the manuscript.

10. How do you understand T in your model? Is this a time of processing or number of iterations?

Response: As mentioned in Section 6, it’s the runtime of the considered algorithms on the considered dataset.

Reviewer 3

1. There are some typos and grammatical errors in Abstract. In addition, the abstract is not attractive. Some sentences in abstract should be modified to make it more attractive for readers.

Response: We thank the reviewer for the suggestion. The Abstract is rewritten to make it more precise and attractive.

2. In Introduction section, it is difficult to understand the novelty of the presented research work. In addition, some references are missing.

Response: In this paper, we proposed a novel CKD prediction model using ensemble learning. We used five different boosting algorithms to check the prediction performance of the model. Along with achieving better results (e.g., accuracy, precision, recall, etc.) and runtime we also assessed the contributions of all the attributes in the dataset that cause CKD.

3. In related work, the existing works about patient disease prediction should be discussed: ‘A smart healthcare monitoring system for heart disease prediction based on ensemble deep learning and feature fusion’, ‘Alzheimer’s disease progression detection model based on an early fusion of cost-effective multimodal data’, Automatic detection of Alzheimer’s disease progression: An efficient information fusion approach with heterogeneous ensemble classifiers’, and ‘An intelligent healthcare monitoring framework using wearable sensors and social networking data.’

Response: We thank the reviewer for suggesting the papers. We really appreciate the works presented in the suggested papers. The papers that are most closely related to our work are cited.

4. The authors should properly select and check the subsection title. There are so many typos and (see section 3 research methodology.

Response: We thank the reviewer for bringing our notice into this. There was a problem with the subsection 3.1 and Section 4 which should be included under Section 3. It is corrected duly. As suggested, a few sections/subsections are renamed.

5. The number given to each section is not correct.

Response: We apologise for the unintentional mistake. The numberings are corrected.

6. Where are the other preprocessing steps? How is the data pre-processed?

Response: The details of the data preprocessing are given in Section 4.4. The steps also pictorially shown in Fig. 1.

7. What about feature selection?

Response: We considered all the featured in the CKD dataset. We assessed the contribution of each feature in CKD. To calculate the feature importance, we used the wrapper method. This is discussed in the manuscript in Section 5.3.3 and also shown in Fig. 13.

8. The results are not properly discussed.

Response: The outcomes of the data preprocessing are presented in Section 4.2. The experimental results are discussed in Section 5.3. The performance of the proposed model is measured using several performance metrics such as accuracy, precision, recall, F1-score, support, AUC-ROC.

9. Captions of the Figures not self-explanatory. The caption of figures should be self-explanatory, and clearly explaining the figure. Extend the description of the mentioned figures to make them self-explanatory.

Response: We thank the reviewer for the suggestion.

Response: Most of the captions are rewritten for better understandability.

10. Equation 2 should be more clearly discussed.

Response: In the original manuscript, there was no Eq. 2.

11. In conclusion section, the future work should be more deeply discussed.

Response: As suggested by the reviewer, the Conclusion section is extended including the limitation of the work and future direction in this domain.

12. The whole manuscript should be thoroughly revised in order to improve its English.

Response: We thank the reviewer for the suggestion. The manuscript is thoroughly checked for English and grammatical mistakes.

Attachment

Submitted filename: Response to Reviewers comments.docx

Decision Letter 1

Muhammad Fazal Ijaz

3 Apr 2023

PONE-D-23-02469R1Chronic Kidney Disease Prediction Using Boosting Techniques based on Clinical ParametersPLOS ONE

Dear Dr. Ganie,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by May 18 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Muhammad Fazal Ijaz

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: No

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: I Don't Know

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: No

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: No

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Authors are recommended to provide more technical data as recommended to make the study evident.

1. Introduction must discuss adequately about the field of study and the limitations of the existing technologies in prediction of chronic kidney diseases.

2. Literature must be tremendously improvised my incorporating some of the relevant studies like https://doi.org/10.3390/s18072183 and https://doi.org/10.3390/diagnostics12123067

3. what addition technical contribution is made other than using the conventional classification techniques like XGBoost, CatBoost, LightGBM, AdaBoost, Gradient boosting.

4. Do the authors have the permission to re-use the data from Apollo Hospitals, Managiri, India. If yes, enclose the permission letter to use the data. (Very important as the dataset is not licensed under CC0: Public Domain.) why authors have not used PIMA dataset.

5. What Data Preprocessing techniques were performed like Normalization or scaling or anything else?

6. The architecture/block diagram of the proposed model must be presented. The notations for a few equations are not discussed in the paragraph above the equation.

7. What are the cases assumed as TP, TN, FP, FN (confusion matrix) in the current study.

8. Authors must provide the details of hyper parameters like training loss, testing loss, training accuracy, and testing accuracy.

9. What was the difference between G.B. and GB... they are interchangeable used across the results section. Check in the confusion matrix figure where G.B. is used and in Figure 8 its GB

10. Table 5 can incorporate some of the recent studies like https://doi.org/10.3390/diagnostics12112739

11. More comparative analysis was desired.

12. There are few typos in the manuscript, authors may crosscheck them.

Reviewer #2: Actually paper is the same, since Authors worked on style and text only and concerns are not solved at all thus i would suggest return to revisions.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Dec 1;18(12):e0295234. doi: 10.1371/journal.pone.0295234.r004

Author response to Decision Letter 1


11 May 2023

Manuscript ID: PONE-D-23-02469R1

Title: Chronic Kidney Disease Prediction Using Boosting Techniques based on Clinical Parameters

We thank the two reviewers for their valuable time on evaluating our manuscript. In this second revision, we have tried our best to address and incorporate these valuable comments. The changes were highlighted by yellow in the revised manuscript.

Response to Reviewer #1

Reviewer #1: Authors are recommended to provide more technical data as recommended to make the study evident.

1. Introduction must discuss adequately about the field of study and the limitations of the existing technologies in prediction of chronic kidney diseases.

Response: We thank the reviewer for this valuable suggestion. In this second revision, the Introduction part is considerably updated as per suggestion. The limitations of existing machine learning algorithms are listed (fourth paragraph). Also, the purpose of our paper is specifically mentioned (last paragraph of Introduction section).

2. Literature must be tremendously improvised my incorporating some of the relevant studies like https://doi.org/10.3390/s18072183 and https://doi.org/10.3390/diagnostics12123067

Response: We appreciate the suggestion by the reviewer. While we felt these two papers either address different problems or their underlying methodology or approaches are completely unrelated. We’ve already incorporated most of the credible papers related to CKD and ensemble learning.

3. what addition technical contribution is made other than using the conventional classification techniques like XGBoost, CatBoost, LightGBM, AdaBoost, Gradient boosting.

Response: We thanks the reviewer for this critical point. In our work, we designed a novel CKD prediction model that includes comprehensive data preprocessing, hyper parameter selection and tuning, feature selection and estimation of feature importance. These cumulatively allowed us to achieve better performance than other prediction models (such as using AdaBoost).

4. Do the authors have the permission to re-use the data from Apollo Hospitals, Managiri, India. If yes, enclose the permission letter to use the data. (Very important as the dataset is not licensed under CC0: Public Domain.) why authors have not used PIMA dataset.

Response: We appreciate the reviewer’s concern on the permission to re-use the considered dataset. We would like to reassure that the dataset is not copyrighted and it is openly available on Kaggle and UCI machine learning repository.

The PIMA dataset comprises the records of diabetes patients. It is not for CKD and hence we did not use this dataset. We respectfully request the reviewer not to ask us to change our research topic in this manuscript, but we are surely happy to extend our work to other disease like diabetes.

5. What Data Preprocessing techniques were performed like Normalization or scaling or anything else?

Response: We thank the reviewer for raising this point. The steps for data preprocessing are included in Section 4.2. In brief, we conducted the following steps as data preprocessing:

a) Identify and replace duplicate values.

b) Identify and replace missing values.

c) Detect and replace the outliers.

d) Convert categorical variables to numerical values using one-hot encoding.

e) Perform data transformation (-1 to 1) and scaling (0 to 1).

6. The architecture/block diagram of the proposed model must be presented. The notations for a few equations are not discussed in the paragraph above the equation.

Response: We thank the reviewer’s point on diagram presentation of our model. The block diagram of the proposed work is given in Fig. 1. We thank the reviewer for pointing the issue of the equation notations. They are duly corrected. We hope the newly revised manuscript satisfies your evaluation.

7. What are the cases assumed as TP, TN, FP, FN (confusion matrix) in the current study.

Response: Thanks for your comment on clarification of these terms. The cases for TP, TN, FP, FN are described in Section 5.3.1. Specifically, as per Fig. 7. The left upper and the right lower boxes denote the number of correct predictions for the patients having (true positive) and not having (true negative) CKD, respectively. The right upper box and the left lower box indicates the number of incorrect predictions for patients having (false positive) and not having (false negative) CKD, respectively.

8. Authors must provide the details of hyper parameters like training loss, testing loss, training accuracy, and testing accuracy.

Response: The details of the hyperparameters are now given in Table 3. Instead of training and testing loss we presented the misclassification rate in Table 4 along with the model accuracy.

9. What was the difference between G.B. and GB… they are interchangeable used across the results section. Check in the confusion matrix figure where G.B. is used and in Figure 8 its GB

Response: We thank the reviewer for pointing out this inconsistency use of the abbreviation. In the second revised manuscript, the GB and other abbreviations are written uniformly throughout the manuscript.

10. Table 5 can incorporate some of the recent studies like https://doi.org/10.3390/diagnostics12112739

Response: Table 5 already includes the available related studies. While we much appreciate this reviewer for valuable suggestion, unfortunately, the suggested paper (AAL and Internet of Medical Things for Monitoring Type-2 Diabetic Patients) is not related to CKD prediction in our work; hence we could not incorporate such studies in Table 5. Please note that we have already cited some other papers suggested by you/or other reviewers in the previous revision, and our manuscript is already long. Thank you again for your kind understanding of scientific merits regarding the scope and integrity.

11. More comparative analysis was desired.

Response: In the previous revision, in response to your comment, we have performed a couple of studies as comparative analysis. While there are many computational works in the biomedical research, it is better for us to focus on CKD prediction. That says, we could not find any more good papers that used Boosting algorithm for CKD prediction. In addition, for some papers we identified in this topic, they were published in the predatory journals or low-level conferences; hence, we did not include them for comparison. We sincerely hope that the reviewer is satisfied with our overall work, and happy to do more if indeed this is required after this petition.

12. There are few typos in the manuscript, authors may crosscheck them.

Response: We thank the reviewer for carefully reading our manuscript. The manuscript is thoroughly rechecked for the typos and grammatical errors. We will also try to work with the typesetters in the proofs stage.

Reviewer #2: Actually paper is the same, since Authors worked on style and text only and concerns are not solved at all thus i would suggest return to revisions.

Response: We thank the reviewer 2 for evaluating our work as a peer reviewer. While we should have done better work in explaining our revision in the previous round, we respectfully disagree with this reviewer’s judgment on our previous revision. We hope there wasn’t any error (e.g., the response letter was not shown) in the manuscript system to this reviewer. In the previous revision, we tried to address every concern of all the three reviewers as practically as possible. The previous responses to reviewers’ comments are given below for your reference. If this reviewer still has the same concern or did not see our response letter, please let us know through editor/editorial office, so that such a misunderstanding would be avoided. Thank you again for your kind service in peer-review system.

Reviewer Reviewer’s concerns (first review) Response/action taken

Reviewer 1 1. The abstract must be re-written, focusing on the technical aspects of the proposed model, the main experimental results, and the metrics used in the evaluation. Briefly discuss how the proposed model is superior We thank the reviewer for the suggestion. The abstract is written including the suggested points.

2. Please make sure the abbreviations are properly used. For example, Area Under Curve-Receiving Operator Characteristic (AUC-ROC). We apologise for the silly mistake. It is duly corrected.

3. Additionally, method names should not be capitalized. Moreover, it is not the best practice to employ abbreviations in the abstract, they should be used when the term is introduced for the first time. We acknowledge the reviewer’s concern regarding capitalisation. However, in the literature the normal convention to write the following terms are:

XGBoost, CatBoost, LightGBM, AdaBoost.

Furthermore, these terms are more popularly known by their abbreviated forms only. Writing the full forms for each of them would increase the word count of the abstract. Nevertheless, as suggested by the reviewer, the full forms of these terms are given in the text when they are used for first time.

4. This paper is not a survey paper to include the references as "[30] [31] [32] [33] [34]." Please provide the significance of each of those studies and add a citation. We thank the reviewer for the suggestion. The goal of the sentence is to establish the importance of ensemble learning in disease diagnosis and treatment. As suggested, we re- written the sentence by including sample research works that addressed different diseases using ensemble learning.

5. Literature must be tremendously improvised for better idea on the field of study and state of art models, authors may consider some of the relevant studies like https://doi.org/10.1371/journal.pone.0271619 We thank the reviewer for the suggestion. We included the suggested paper in the related work. We also improved the literature survey following the paper. A couple of recent papers are included in the related work.

6. Boosting algorithms must be adequately explain, for better idea refer and include https://doi.org/10.3390/healthcare10010085 We thank the reviewer for the suggestion. The boosting algorithms are elaborated with suitable mathematical equations.

7. How are the feature weights calculated, what was the approach or the technique that is being used. To calculate the feature importance, we used the wrapper method. This is included in the manuscript with suitable reference.

8. Manuscript is having too many sub-sections, please minimize for better readability of the study. We thank the reviewer for bringing our notice into this. There was a problem with the subsection 3.1 and Section 4 which should be included under Section 3. It is corrected duly.

We structured the manuscript into different subsections which allowed us to express the non-overlapping contents precisely and exclusively. This would help the readers to focus and understand the fragments of the experiment in a granular way.

9. The section Data Standardization and Normalisation must be discussed as data pre-processing phase, that has to be in background section of the manuscript. As suggested, the section Data Standardization and Normalisation is included in the Data Pre-processing section.

10. What are the cases that are assumed as TP, TN, FP and FN, please explain them clearly, for better idea refer https://doi.org/10.3390/s21082852 We thank the reviewer for the suggestion. Each prediction cases are stated with respect to Figure 7.

11. Authors may present the loss functions for better comprehensibility of each of the models used in the proposed model. For better idea refer https://doi.org/10.3390/s21165386 The misclassification rate of incorrect prediction is given in Table 4.

12. Majority of the figures lack the clarity, they quality is fair but they must be explained in the text and the figures must be cited. The figures are described and cited in the text.

13. More comparative analysis with state-of-art models is desired. As suggested by the reviewer, our work has been compared with a couple of papers in addition to the existing 6 papers, as shown in Table 5.

14. By considering the current form of the conclusion section, it is hard to understand by PLos One Journal readers. It should be extended with new sentences about the necessity and contributions of the study by considering the authors' opinions about the experimental results derived from some other well-known objective evaluation values if it is possible The necessity and contribution of this research work is included in the Conclusion section.

15. English proofreading is strongly recommended for a better understanding of the study, and the quality of the figures must be tremendously improved. The manuscript is proofread for possible grammatical and writing mistakes.

Regarding the figures, most of them are program generated and large in size. Accommodating in a smaller scale makes a couple of figures look unclear. However, they are perfectly readable by zooming.

Reviewer 2 1. What are main aspects of novelty and advances in the described ideas? In this paper, we proposed a novel CKD prediction model using ensemble learning. We used five different boosting algorithms to check the prediction performance of the model. Along with achieving better results (e.g., accuracy, precision, recall, etc.) and runtime we also assessed the contributions of all the attributes in the dataset that cause CKD.

2. What are limitations of presented approaches? How the models work in different scenarios of operation? What are weak points of presented ideas? We thank the reviewer for the suggestion. The limitations of the work in included in the Conclusion section.

3. Related ideas: Deep neural network correlation learning mechanism for CT brain tumor detection, BiLSTM deep neural network model for imbalanced medical data of IoT systems, Deep learning for neurodegenerative disorder (2016 to 2022): A systematic review. We thank the reviewer for the valuable suggestion. We’ve already started working on the deep learning based models for different disease predictions.

4. Compare model to other in different operation and in different positioning of the input data. We tried with different combinations of the data preprocessing, feature selection and hyperparameter tuning and came out with the best performing model. The model was tried with five boosting algorithms. Finally, the performance of the best working combination of the proposed model with AdaBoost is compared with a number of state-of-the-art findings.

5. What are future trends in the development of this type systems? How the development would work in different configurations? What kind of transfer and network configurations are necessary for your model? In future, more sophisticated disease prediction models will be prevalent in the medical diagnosis and treatment.

The proposed model can be used for other healthcare datasets that share the commonality of features.

The configuration setup will depend on the particular application requirement and the properties of the available dataset.

The future works of this work is mentioned in the Conclusion section.

6. How to set optimal coefficients for these models? Did you test other configurations? How were these selected? We tested the performance of the prediction model for combination of different coefficient values for all the tuneable parameters. Among them, the optimal value sets were selected, as shown in Table 3.

7. Did you test the option to transform the knowledge before processing? We carried out the background work to set up the base prediction model and tested it on different ensemble algorithms. Among them, in the given setup, AdaBoost performed best.

8. There are no comparisons to other models so we are not able to see advances of your processing. The proposed model is compared with several similar published works, as shown in Table 5. It can be observed that our model outperforms the other compared works.

9. Your fig 1 is not much informative since there are no details on your model thus we are not able to repeat your experiment. Fig. 1 shows only the overall flow of the paper. The details of each step are elaborately discussed in the manuscript.

10. How do you understand T in your model? Is this a time of processing or number of iterations? As mentioned in Section 6, it’s the runtime of the considered algorithms on the considered dataset.

Reviewer 3 1. There are some typos and grammatical errors in Abstract. In addition, the abstract is not attractive. Some sentences in abstract should be modified to make it more attractive for readers. We thank the reviewer for the suggestion. The Abstract is rewritten to make it more precise and attractive.

2. In Introduction section, it is difficult to understand the novelty of the presented research work. In addition, some references are missing. In this paper, we proposed a novel CKD prediction model using ensemble learning. We used five different boosting algorithms to check the prediction performance of the model. Along with achieving better results (e.g., accuracy, precision, recall, etc.) and runtime we also assessed the contributions of all the attributes in the dataset that cause CKD.

3. In related work, the existing works about patient disease prediction should be discussed: ‘A smart healthcare monitoring system for heart disease prediction based on ensemble deep learning and feature fusion’, ‘Alzheimer’s disease progression detection model based on an early fusion of cost-effective multimodal data’, Automatic detection of Alzheimer’s disease progression: An efficient information fusion approach with heterogeneous ensemble classifiers’, and ‘An intelligent healthcare monitoring framework using wearable sensors and social networking data.’ We thank the reviewer for suggesting the papers. We really appreciate the works presented in the suggested papers. The papers that are most closely related to our work are cited.

4. The authors should properly select and check the subsection title. There are so many typos and (see section 3 research methodology. We thank the reviewer for bringing our notice into this. There was a problem with the subsection 3.1 and Section 4 which should be included under Section 3. It is corrected duly. As suggested, a few sections/subsections are renamed.

5. The number given to each section is not correct. We apologise for the unintentional mistake. The numberings are corrected.

6. Where are the other preprocessing steps? How is the data pre-processed? The details of the data preprocessing are given in Section 4.4. The steps also pictorially shown in Fig. 1.

7. What about feature selection? We considered all the featured in the CKD dataset. We assessed the contribution of each feature in CKD. To calculate the feature importance, we used the wrapper method. This is discussed in the manuscript in Section 5.3.3 and also shown in Fig. 13.

8. The results are not properly discussed. The outcomes of the data preprocessing are presented in Section 4.2. The experimental results are discussed in Section 5.3. The performance of the proposed model is measured using several performance metrics such as accuracy, precision, recall, F1-score, support, AUC-ROC.

9. Captions of the Figures not self-explanatory. The caption of figures should be self-explanatory, and clearly explaining the figure. Extend the description of the mentioned figures to make them self-explanatory. We thank the reviewer for the suggestion. Most of the captions are rewritten for better understandability.

10. Equation 2 should be more clearly discussed. There was no Eq. 2 in our original manuscript (first submission). If the reviewer refers to other equation, please specify and we will add some more detail, or check for this issue.

11. In conclusion section, the future work should be more deeply discussed. As suggested by the reviewer, the Conclusion section is extended including the limitation of the work and future direction in this domain.

12. The whole manuscript should be thoroughly revised in order to improve its English. We thank the reviewer for the suggestion. The manuscript is thoroughly checked for English and grammatical mistakes.

Attachment

Submitted filename: Response to Reviewers comments (2nd review).docx

Decision Letter 2

Anwar PP Abdul Majeed

7 Jul 2023

PONE-D-23-02469R2Chronic Kidney Disease Prediction Using Boosting Techniques based on Clinical ParametersPLOS ONE

Dear Dr. Ganie,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Aug 21 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Anwar P.P. Abdul Majeed

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

Reviewer #4: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: No

Reviewer #4: Partly

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: No

Reviewer #4: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

Reviewer #4: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: No

Reviewer #4: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors have addressed all the recommendations of the reviewers in a reasonable manner, manuscript in the current from may be considered for the further phase of the editorial process.

-Captions of the Figures not self-explanatory. The caption of figures should be self-explanatory, and clearly explaining the figure. Extend the description of the mentioned figures to make them self-explanatory.

- The corresponding code may be enclosed to make the study evident.

Reviewer #2: Papers is presented in the same way as visible even after presented changes, thus i confirm my opinion from previous round

Reviewer #4: Dear Authors,

Hereby is my initial comments:

1. The dataset used is a public dataset. You need to compare related works within the similar dataset instead of comparing with other datasets. To enhance your novelty.

2. The dataset is an imbalance dataset with ratio 150:250 for with/without CKD. In my opinion, you should make sure the dataset is imbalanced or else you need to use a proper performance measure for the imbalanced dataset. By having an imbalanced dataset, your results are biased results! See your results, CKD and non CKD have big gaps.

3. Did you select important features before doing hyperparameters tuning? How many features are discarded and how many features are retained? If so, the explanation of feature importance should come before hyperparameters tuning.

4. In standard ML procedure, important features are selected to be fed into the classifiers to improve the accuracy of the model.

5. The wrapper method is used for feature importance. What type of wrapper method is used, Forward selection, Backward elimination, or Bi-directional elimination (Stepwise Selection)? Kindly explain.

6. What are the initial parameters for each boosting algorithm before doing hyperparameters tuning? Table 3 is a found best parameters not hyperparameter tuning. Revise the table caption.

7. Which dataset is used for hyperparameters tuning? 60% of the train set? What cross validation is used for this tuning?

8. What method is used for hyperparameters tuning? Grid search method for all algorithms? Kindly include the explanation of the grid search method.

9. The accuracy results should have both for train and test set. Figure 8 is a result for which set? The obtained confusion matrix in Figure 7 is for the test set?

10. Kindly explain what happened in this stage c) Detect and replace the outliers. Why did you replace the outlier? The outlier is your real data.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #4: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Dec 1;18(12):e0295234. doi: 10.1371/journal.pone.0295234.r006

Author response to Decision Letter 2


24 Jul 2023

Response to Reviewers’ comments

Reviewer #1: The authors have addressed all the recommendations of the reviewers in a reasonable manner, manuscript in the current from may be considered for the further phase of the editorial process.

-Captions of the Figures not self-explanatory. The caption of figures should be self-explanatory, and clearly explaining the figure. Extend the description of the mentioned figures to make them self-explanatory.

Response: We thank the reviewer for the suggestion. We renamed the figures and tables where required.

- The corresponding code may be enclosed to make the study evident.

Response: We agree with the reviewer’s suggestion. We plan to compile the code onto an open-access repository such as GitHub or our website soon.

Reviewer #2: Papers is presented in the same way as visible even after presented changes, thus i confirm my opinion from previous round

Response: In our communication with the editor, we will not need to respond this reviewer’s criticism because he/she continuously ignored what we revised our previous versions of manuscript in response to his/her comments.

Reviewer #4: Dear Authors, hereby is my initial comments:

1. The dataset used is a public dataset. You need to compare related works within the similar dataset instead of comparing with other datasets. To enhance your novelty.

Response: We appreciate the suggestion made by the reviewer. We compared our method with similar work that used the same dataset, i.e., the Chronic Kidney Dataset collected from the UCI machine learning repository. The comparison summary is given in Table 5. In the related work section, to discuss the state-of-the-art on chronic kidney disease prediction, we needed to explore all the recent work on the same topic.

2. The dataset is an imbalance dataset with ratio 150:250 for with/without CKD. In my opinion, you should make sure the dataset is imbalanced or else you need to use a proper performance measure for the imbalanced dataset. By having an imbalanced dataset, your results are biased results! See your results, CKD and non CKD have big gaps.

Response: We sincerely thank the reviewer for highlighting this omission. We’ve reworked with the dataset to make it balanced, as shown in Fig. 2. The whole experiment is conducted on this updated balanced dataset, ensuring that there are no biases in the outcome.

3. Did you select important features before doing hyperparameters tuning? How many features are discarded and how many features are retained? If so, the explanation of feature importance should come before hyperparameters tuning.

Response: We thank the reviewer for raising this query. In the revised manuscript, we eliminated the insignificant/non-contributing features from the dataset after finding the important features. Altogether nine attributes (‘ane’, ‘appet’, ‘ba’, ‘cad’, ‘pc’, ‘pcc’, ‘pe’, ‘su’, and ‘wc’) were discarded. We repeated the experiment with the modified dataset, the details of which are discussed in sections 5.2 to 5.5.

4. In standard ML procedure, important features are selected to be fed into the classifiers to improve the accuracy of the model.

Response: The reviewer has rightly stated that feeding the important features into the prediction model improves its performance. Complying with this suggestion, we repeated the experiment with the modified dataset (after eliminating the non-contributing features), the details of which is discussed in sections 5.2 to 5.5.

5. The wrapper method is used for feature importance. What type of wrapper method is used, Forward selection, Backward elimination, or Bi-directional elimination (Stepwise Selection)? Kindly explain.

Response: We thank the reviewer for mentioning this missed out point. We used forward selection as the wrapper method. It is now mentioned in the manuscript.

6. What are the initial parameters for each boosting algorithm before doing hyperparameters tuning? Table 3 is a found best parameters not hyperparameter tuning. Revise the table caption.

Response: We thank the reviewer for identifying the unintentional mistake. We tried with different values of the considered parameters. The best values found for each parameter are shown in Table 3. The caption of the table is changed as suggested.

7. Which dataset is used for hyperparameters tuning? 60% of the train set? What cross validation is used for this tuning?

Response: The hyperparameter tuning was done on the training set (i.e., 60%). We applied k-fold (k=6) cross-validation during the training phase itself. A separate subsection (5.4) is added in the manuscript to cover the validation.

8. What method is used for hyperparameters tuning? Grid search method for all algorithms? Kindly include the explanation of the grid search method.

Response: Besides the grid search method, we also tried with random search method but found that the grid search method provided the best results. Also, we preferred the grid search method because it is used in most of the literature for hyperparameter tuning in disease prediction. The further justification for using the grid search method is discussed in Section 5.3.

9. The accuracy results should have both for the train and test set. Figure 8 is a result for which set? The obtained confusion matrix in Figure 7 is for the test set?

Response: We thank the reviewer for the suggestion. We have added the training accuracy along with the testing accuracy in Fig. 8 (now Fig. 9). The confusion matrix shown in Fig. 7 (now Fig. 8) is of only the test set. Adding a confusion matrix of the training set would occupy another whole page. To keep the manuscript length standard, we decided not to add in the main text. If this reviewer still prefers this, we can do next time.

10. Kindly explain what happened in this stage c) Detect and replace the outliers. Why did you replace the outlier? The outlier is your real data.

Response: We thank the reviewer for this valuable point. Outliers are data points that considerably differ from the majority of the data in the dataset. They are generally resulted due to various errors. Outliers can significantly affect how well machine learning models function. They can skew the data distribution, making it non-normal or non-uniform. Therefore, it is important to deal with them. Removing or replacing outliers can make prediction models more robust and less susceptible to extreme values, allowing them to capture the underlying patterns more accurately.

Attachment

Submitted filename: Response to Reviewers comments (3rd review).docx

Decision Letter 3

Anwar PP Abdul Majeed

20 Nov 2023

Chronic Kidney Disease Prediction Using Boosting Techniques based on Clinical Parameters

PONE-D-23-02469R3

Dear Dr. Ganie,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Anwar P.P. Abdul Majeed

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #4: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #4: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #4: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #4: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #4: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #4: All of the comments (10 comments) have been well addressed by the authors. Overall, I am satisfied with the revised article. The article is ready to be published.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #4: No

**********

Acceptance letter

Anwar PP Abdul Majeed

23 Nov 2023

PONE-D-23-02469R3

Chronic Kidney Disease Prediction Using Boosting Techniques based on Clinical Parameters

Dear Dr. Ganie:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Anwar P.P. Abdul Majeed

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response to Reviewers comments.docx

    Attachment

    Submitted filename: Response to Reviewers comments (2nd review).docx

    Attachment

    Submitted filename: Response to Reviewers comments (3rd review).docx

    Data Availability Statement

    All relevant data can be found at https://archive.ics.uci.edu/ml/datasets/chronic_kidney_disease.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES