Skip to main content
Journal of Healthcare Engineering logoLink to Journal of Healthcare Engineering
. 2021 Feb 20;2021:6247652. doi: 10.1155/2021/6247652

Key Experimental Factors of Machine Learning-Based Identification of Surgery Cancellations

Fengyi Zhang 1,2, Xinyuan Cui 2,3, Renrong Gong 4, Chuan Zhang 5,, Zhigao Liao 1
PMCID: PMC7914093  PMID: 33688420

Abstract

This study aimed to provide effective methods for the identification of surgeries with high cancellation risk based on machine learning models and analyze the key factors that affect the identification performance. The data covered the period from January 1, 2013, to December 31, 2014, at West China Hospital in China, which focus on elective urologic surgeries. All surgeries were scheduled one day in advance, and all cancellations were of institutional resource- and capacity-related types. Feature selection strategies, machine learning models, and sampling methods are the most discussed topic in general machine learning researches and have a direct impact on the performance of machine learning models. Hence, they were considered to systematically generate complete schemes in machine learning-based identification of surgery cancellations. The results proved the feasibility and robustness of identifying surgeries with high cancellation risk, with the considerable maximum of area under the curve (AUC) (0.7199) for random forest model with original sampling using backward selection strategy. In addition, one-side Delong test and sum of square error analysis were conducted to measure the effects of feature selection strategy, machine learning model, and sampling method on the identification of surgeries with high cancellation risk, and the selection of machine learning model was identified as the key factors that affect the identification of surgeries with high cancellation risk. This study offers methodology and insights for identifying the key experimental factors for identifying surgery cancellations, and it is helpful to further research on machine learning-based identification of surgeries with high cancellation risk.

1. Introduction

Surgery cancellation is a well-recognized health care quality problem that harms patients and wastes resources, leading to considerable losses for medical institutes and the entire health care system [13]. It forces scarce operative resources to remain idle and hinders patients' access to operative services [4]. Estimates of this revenue loss range between USD 1,430 and USD 1,700 for each cancelled case in US hospitals not on a fixed annual budget [5, 6]. In a review of surgery cancellations worldwide [4], the global cancellation rate (CR) generally ranges from 4.65% to 30.3%, which is a high proportion that urgently needs to be reduced. Identification of surgeries with high cancellation risk could provide information for health care service management and enable the adoption of preventive actions for achieving a lower CR [4]. Hence, it is of great value to identify surgeries with high cancellation risk.

For reflecting relevant information on patients and medical operation institutions, the hospital information system (HIS) plays an important role in health care service management, including surgery scheduling. Nowadays, health care service management-related studies on HISs have been conducted in many important fields [716], such as research on hospital admission rates [7, 8], clinical medication rules [9, 10], and referral management in hospitals [11, 12]. Particularly, the study of HIS in the field of identification of surgeries with high cancellation risk has proved that applying HIS data to identify surgeries with high and low risks of cancellation is feasible [4].

To date, most surgery cancellation studies have mainly focused on the predictors or risk factors of surgery cancellation. In the risk-factor-based category of cancellation, surgery cancellations generally include institutional resource-, capacity-, and patient-related cancellations. A significant percentage of surgery cancellations could be avoided, according to risk-factor-based analysis [1719]. The risk factors of surgery cancellation include patient, admission, workup, and surgery schedule information, as well as administrative issues and surgery process records [4, 17, 2026].

In recent years, various studies have focused on reducing surgery cancellation [21, 2730]. For example, a team of researchers [27] used discrete event simulation modeling to represent perioperative processes and tested the effects of three scenarios on the number of surgical cancellations. Another team [28] simulated an anesthesiology preoperative assessment clinic to quantify the impact of patient information deficiency to mitigate the problem of surgery delay or cancellation. These studies used industrial engineering techniques to investigate means for reducing the number of surgical cancellations across the system but did not focus on identifying surgery with high cancellation risk. Among the works focusing on identifying surgeries with high cancellation risk, a retrospective cohort study [29] examined the association between patient, surgeon, and system factors and proved that several patient and system factors can be used to identify surgeries with a high likelihood of cancellation. The factors associated with surgery cancellation have been evaluated using chi-squared tests and multivariate logistic regression analyses [21]. Using multilevel logistic regression, an observational cohort study [30] identified patient- and hospital-level factors associated with cancellation owing to inadequate bed capacity. However, the accuracy of the above-mentioned studies only using traditional statistics models is low [29].

Machine learning is a powerful and effective tool for medical study. Machine learning has seen many applications in the fields of health care management [3133], health care cost prediction [34, 35], and health care insurance [3638]. Various machine learning models, which are of better performance compared with traditional statistics models, have been used in the field of identification of surgeries with high cancellation risk as well [4]. Particularly, Luo et al. [4] used sampling methods to handle the imbalance of the distribution of cancellation. Considering the performance of feature selection used in health care services [39, 40], Liu et al. [41] developed predictive models of last-minute surgery cancellation, in which forward selection was used as a feature selection method. However, there is no research that measured the effects of feature selection strategy, machine learning model, and sampling methods on the identification of surgery with high cancellation risk and identified the key factors of it.

This study aimed to provide effective methods for the identification of surgeries with high cancellation risk based on machine learning models and analyzed the key factors that affect the identification performance. One-side Delong test and sum of square error analysis were conducted to measure the effects of feature selection strategy, machine learning model, and sampling method on the identification of surgeries with high cancellation risk. This study offers methodology and insights for identifying the key experimental factors for identifying surgery cancellations, and it is useful to further research on machine learning-based identification of surgeries with high cancellation risk, in designing experimental process.

The rest of the paper is organized as follows. Section 2 provides detailed information about data collection and preparation, the methods used in this study, and model setup. The third part summarizes the experimental results. Section 4 further discusses the experiment and presents the findings. In Section 5, we draw conclusions from the findings and point out the direction of future work.

2. Data and Methods

2.1. Data

The data of this study were based on HIS sourced from West China Hospital, which is the largest hospital in southwest China. There are 70 operation rooms in West China Hospital nowadays, most of them usually open from 8 a.m. to 8 p.m., and daily average opening hours reach 12 hours. Overall, the data contained 5,125 cases from January 1, 2013, to December 31, 2014, of which 810 were cancelled (positive) and 4,315 were not, providing a CR of 15.80%. The hospital implemented the surgery day system. In this system, the surgeon has main surgery days; that is, the surgeon has the priority to use the operating room and its equipment on these days. All surgeries were scheduled one day in advance, and after confirmation, medical staff (surgeon, anesthesiologist, and nurse), the operation room, and the patient are bundled together. All cancellations were of institutional resource- and capacity-related types. Apart from the features originally obtained from HIS, we designed some features, according to the experience and knowledge of senior health care managers in West China Hospital, and they are whether the surgeon had surgery before (WSHSB), whether the surgery day was a legal holiday (WSDLH), whether it was the main surgery day (WMSD), whether surgeries have been cancelled (WC), and the number of days admitted (NDA). All the collected features of surgeries are listed in Table 1.

Table 1.

Predictors collected in this study.

Category Number of predictors Predictor(s)
Patients' information 4 Name, age, sex, and AT
Admission information 4 NDA, visit number, identification number of patient, and register number
Workup information 7 Drug allergy, names of drugs administered, blood type, WHSB, SN, and ST
Surgery schedule information 7 ONS, OR, surgery date, surgery time, surgeon, NSOD, WSHSB, and purpose of surgery
Administrative issues 10 Operation staff, department, ward, BD, last updated time, staff who last updated the information, WSDLH, WMSD, WC, and surgery expenditure
Surgery process records 23 Actual date/time when surgery began/ended, actual date/time when patient left OR, actual date when anesthesia was started, actual time when anesthesia was ended, actual date/time when predictive medicine was administered, body temperature, blood transfusion in surgery, autologous blood, allogeneic blood, plasma, thrombocyte, pathological examination, state of consciousness, general skin conditions, special skin conditions, drainage situation, surgery item delivery, anesthesia degree, and surgical incision category

AT: anesthesia type. BD: bed number. NDA: number of days admitted. NSOD: number of surgeries in the OR on the day. ONS: order number of surgery. OR: operating room. SN: surgery name. ST: surgery type. WC: whether surgery is cancelled. WHSB: whether there has been a surgery before. WMSD: whether it is the main surgery day. WSDLH: whether the surgery day is a legal holiday. WSHSB: whether the surgeon has surgery before.

In the following experiment process, we considered not only the relationship between predictors and surgery cancellations but also the accessibility of predictors. As a result, 14 related predictors (features) were preliminarily refined, as described in Table 2, which covered all five information categories mentioned above. Among them, surgery schedule information contained five predictors: operating room (OR), surgeon, number of surgeries in the OR on a day (NSOD), the order number of surgery (ONS), and WSHSB. Then, patients' information and administrative issues contained three predictors each. For patients' information, they were age, sex, and anesthesia type (AT). For administrative issues, they were WSDLH, WMSD, and WC. Subsequently, workup and admission information contained two predictors and one predictor, respectively. Workup information contained WHSB and surgery type (ST), whereas admission information contained NDA.

Table 2.

Predictors considered in this study.

Category Number of predictors Predictor(s)
Patients' information 3 Age, sex and AT
Admission information 1 NDA
Workup information 2 WHSB and ST
Surgery schedule information 5 OR, surgeon, NSOD, ONS, and WSHSB
Administrative issues 3 WSDLH, WMSD, and WC

AT: anesthesia type. NDA: number of days admitted. NSOD: number of surgeries in the OR on the day. ONS: order number of surgery. OR: operating room. ST: surgery type. WC: whether surgery is cancelled. WHSB: whether there has been a surgery before. WMSD: whether it is the main surgery day. WSDLH: whether the surgery day is a legal holiday. WSHSB: whether the surgeon has surgery before. Predictor(s): predictors mentioned above were preliminarily identified through expert interviews.

2.2. Methods

The methods used in this study are introduced in the following three aspects: feature selection strategies, machine learning models, and sampling methods.

2.2.1. Machine Learning Models

Considering that the objective of this study is to identify surgery cancellation based on historical HIS data and different predictors, we modeled it as a supervised classification problem and utilized a representative set of machine learning models: random forest (RF), logistic regression (LR), extreme gradient boosting-tree (XGBoost-tree), support vector machine-linear (SVM-linear), and neural networks (NNET).

RF is a classifier composed of an ensemble of decision trees for training and predicting, which is widely used in medical management [4244]. It is known for the high predictive performance and ability to find complex interactions among features [41, 45]. LR, a classification algorithm derived from linear regression, is a common approach employed in prediction and reasonable benchmark for evaluating other models. Compared with SVM and NNET, it has better interpretability that is important for model understanding and interpretation. XGBoost is an improved algorithm based on the gradient boosting decision tree with more detailed classification, XGB-Linear and XGB-tree. As for XGB-tree, it can construct boosted trees efficiently and when performing node splitting, the gain calculation of different predictors can be performed in parallel. SVM-linear is a kind of generalized linear classifier. Because of its advantages of solving high-dimensional pattern recognition problems and high accuracy, it is applied in this study. NNET is a model like the human brain's ability to predict and categorize, which learns the relationship between independent variables and dependent variables. Nowadays, it has been successfully applied in the classification and prediction of biological and medical data [4648].

2.2.2. Feature Selection Strategies

This study takes four commonly used feature selection strategies into consideration, and they were forward selection strategy, backward selection strategy, LASSO-based strategy, and importance-based strategy. Forward selection starts with an empty set and iteratively adds the most important feature to the target feature subset from the candidate feature set, while backward selection iteratively removes the least important feature from the candidate feature set [49]. LASSO-based strategy refers to the cost function of the linear regression model added with the constraint of the L1 norm. It uses the control parameter for variable selection and complexity adjustment and is widely used in the medical field [50]. As for the importance-based strategy, RF, an excellent classifier model that has good applicability for feature selection [51, 52], was used to generate a dataset of the importance of each feature after training, and features with positive importance were considered as the most useful features to model training. Hence, feature selection using RF was conducted to select the useful predictors for the identification of the surgeries with high cancellation risk.

2.2.3. Sampling Methods

This study involves three sampling methods: oversampling, undersampling, and original sampling. Oversampling was intended to extract negative samples with replacement until the number of them was consistent with the number of positive samples. Conversely, undersampling extracted positive samples without replacement until their number was the same as the negative samples. Both methods change the class distribution of training data and have been used to address class imbalance [53]. These methods have performed well in several fields, such as in churning predictions related to bank credit cards [54] and classifying poor households [55]. Meanwhile, original sampling is the sampling of the original data set without making changes.

2.3. Experiment Setup

Figure 1 shows the entire experimental process mentioned below. For the preliminarily determined 14 related predictors, they may not certainly lead to the best performance of identification of surgery cancellation. Hence, the four feature selection strategies mentioned above (forward selection strategy, backward selection strategy, LASSO-based strategy, and importance-based strategy) were considered to achieve better performance. Subsequently, for a certain strategy, N predictors were obtained. To a large extent, the value of N is different for each strategy. Hence, predictors selected by original strategy (i.e., no feature selection is conducted) and four different feature selection strategies were applied to the following experiments, respectively.

Figure 1.

Figure 1

Experiment flowchart.

All samples were divided into two sets, the train and test sets, at a ratio of 8 : 2. Based on this division, the train set was divided into the actual train and the validation sets, at a ratio of 6 : 2. Because of the imbalance in the positive-negative ratio (2 : 11) of the actual train set, we employed not only the original sampling but also over- and undersampling to achieve better performance. The train set and the actual train set were used to train the machine learning models; the validation set was used to determine the hyperparameters of models; and the test set was employed to validate the performance of the machine learning models. For each model, we performed a fivefold cross validation against the test set, nested within which was a fourfold cross validation against the validation set. Cross-validation methods are used to generate folds randomly, which refer to the combination of training and test data subset splits for training and validating machine learning models [56].

In this study, we designed 75 schemes (i.e., 5 × 5 × 3; five feature selection strategies, five machine learning models, and three sampling methods correspond to each other, as mentioned above), and each scheme was run to obtain the performance metrics of the test set. Fivefold cross validation was performed for each scheme to find and validate the optimal model with the best performance. Hence, 375 results of experiments were obtained eventually, from which the scheme with the best performance could be found and the influence of the factors (i.e., feature selection strategy, machine learning model, and sampling method) in surgery cancellation forecasting could be evaluated.

As the identification of surgeries with high cancellation risk in this study belongs to the binary classification problem, the performance of identification was measured according to seven metrics: accuracy, sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), net reclassification index (NRI), and area under the curve (AUC) of the receiver operating characteristic (ROC). Sensitivity, specificity, PPV, and NPV are the metrics for model evaluation, which are used to reflect performance in a certain aspect. Sensitivity refers to the ratio of the correctly predicted positive sample number to the total number of true positive samples; conversely, specificity refers to the ratio of correctly predicted negative samples to the total number of true negative samples. Meanwhile, both PPV and NPV are metrics that focus on predictive samples. PPV refers to the ratio of the number of correctly predicted positive samples to the number of predicted positive samples, whereas NPV refers to the ratio of the number of correctly predicted negative samples to the number of predicted negative samples. Accuracy, NRI, and AUC are all used to reflect the overall performance of the model. Accuracy is the ratio of the correctly predicted sample number to the total predicted sample number. It does not distinguish the predicted sample as positive or negative. NRI is a measure of the change in risk prediction obtained when the risk marker under evaluation is added to an existing risk prediction model [57]. It was intended to serve as a summary measure to highlight the difference between two models [58]. AUC considers the imbalance of positive and negative samples and is often used with the ROC curve to illustrate performance assessments so that sensitivity and specificity can be considered in a comprehensive manner. In addition, feasibility is defined as the ability to make a considerable identification, and robustness refers to the measured performance close to the essential performance. In this study, AUC was considered to be the key metric; feasibility and robustness both focus on AUC.

The mentioned metrics were firstly analyzed for top schemes, and then we analyzed statistics (mean, maximum, minimum, etc.) of all schemes. Considering AUC being the key metric, we summarized the average AUCs grouped by different factors (feature selection strategy, machine learning model, and sampling method). We also conducted a Delong test [59] to evaluate the impact of different methods on predicting results. The variables involved in the study include feature selection methods, machine learning methods, sampling methods, and fivefold cross validation. Hence, the difference in AUC between specific methods can be evaluated by controlling other variables unchanged. P < 0.05 is considered statistically significant. In addition, the between-groups sum of squares (BGSS), within-groups sum of squares (WGSS), and total sum of squares (TSS) were used to measure deviation of the AUCs grouped by machine learning models, sampling methods, and feature selection strategies, and their definitions are given as follows:

BGSS=i=1rnix¯ix¯2,WGSS=i=1rj=1nix¯ixij2,TSS=i=1rj=1nix¯xij2=WGSS+BGSS, (1)

where r refers to the number of groups, ni refers to the number of examples belonging to group i, x¯ refers to the mean of all samples, x¯i refers to the mean of ith group, and xij refers to the jth sample of the ith group. Given certain TSS, the larger the GBSS (the less the WGSS), the better the grouping. The experiments implemented were based on R software (version 3.61); the identification was conducted with the “caret” packages.

3. Results

We analyzed the experimental results in both scheme and factor level.

3.1. Analysis at Scheme Level

The schemes were mainly measured by averaging mentioned metrics of the fivefold cross validation, and the top 15 schemes in the test set are shown in Table 3, in descending order of average AUC. A model is considered as considerable predictive performance, if it is of a higher than 0.7 AUC [60, 61]. All NRIs were measured compared to the scheme of RF, backward selection strategy, and original sampling. Table 3 indicates the following: (1) The RF model with original sampling using backward selection strategy achieved the best performance according to accuracy (0.8578) and AUC (0.7199). (2) All top 9 schemes were of RF models, and RF model accounted for 11 of the top 15. Meanwhile, the schemes with a higher than 0.7 AUC were all RF models. (3) For the sampling method, original sampling and oversampling both accounted for 6 in the top 15 schemes, and the top 4 were original sampling. For the schemes with a higher than 0.7 AUC, both oversampling and original sampling accounted for half of all (4) For the NRI, there were 12 schemes with negative results, indicating that their performance was worse than the first one. Only two schemes achieved the positive NRI results, but differences were quite small (RF, original strategy, and original sampling: 0.0020; RF, importance-based strategy, and original sampling: 0.0057).

Table 3.

Top 15 schemes in the test set.

Model Strategy Sampling AUC Accuracy Sensitivity Specificity PPV NPV NRI
1 RF Backward selection strategy Original sampling 0.7199 0.8578 0.6712 0.8668 0.1963 0.9819 0.0000
2 RF Original strategy Original sampling 0.7136 0.8574 0.6740 0.8660 0.1901 0.9826 0.0020
3 RF Forward selection strategy Original sampling 0.7135 0.8560 0.6284 0.8692 0.2173 0.9759 −0.0404
4 RF Importance-based strategy Original sampling 0.7131 0.8574 0.6783 0.8654 0.1852 0.9835 0.0057
5 RF Backward selection strategy Oversampling 0.7055 0.8343 0.4579 0.8722 0.2654 0.9411 −0.2079
6 RF Forward selection strategy Oversampling 0.7034 0.8154 0.3904 0.8736 0.2963 0.9129 −0.2740
7 RF Importance-based strategy Oversampling 0.7029 0.8398 0.4869 0.8720 0.2580 0.9490 −0.1791
8 RF Original strategy Oversampling 0.7018 0.8394 0.4874 0.8726 0.2630 0.9476 −0.1780
9 RF Forward selection strategy Undersampling 0.6847 0.6176 0.2369 0.9003 0.6383 0.6137 −0.4008
10 NNET LASSO-based strategy Original sampling 0.6814 0.8398 0.4167 0.8424 0.0123 0.9934 −0.2789
11 RF Backward selection strategy Undersampling 0.6814 0.6328 0.2436 0.9009 0.6284 0.6336 −0.3935
12 RF Original strategy Undersampling 0.6808 0.6310 0.2384 0.8960 0.6074 0.6355 −0.4036
13 SVM-linear LASSO-based strategy Oversampling 0.6797 0.6597 0.2523 0.8970 0.5877 0.6732 −0.3887
14 LR LASSO-based strategy Oversampling 0.6794 0.6457 0.2493 0.9012 0.6185 0.6508 −0.3875
15 LR LASSO-based strategy Original sampling 0.6793 0.8429 0.5492 0.8459 0.0346 0.9947 −0.1429

LR: logistic regression model. NNET: neural networks. RF: random forest. SVM-linear: support vector machine-linear. XGBoost-tree: extreme gradient boosting-tree. Accuracy: ratio of the correctly predicted sample number to the total predicted sample number. Sensitivity: ratio of correctly predicted positive samples to the total number of true positive samples. Specificity: ratio of correctly predicted negative sample number to the total number of true negative samples. PPV: positive predictive value. NPV: negative predictive value. AUC: area under the receiver operating characteristic curve. NRI: net reclassification index; all NRIs were measured compared to the scheme of RF, backward selection strategy, and original sampling. The values in the table are averages of the fivefold cross validation.

In addition, the statistics of each metric over the 75 schemes are shown in Table 4. According to Table 4, we can find the following: (1) for all the schemes, the specificity and NPV were quite high, with an average value of more than 0.75 (specificity: 0.8751, NPV: 0.7760), and the maximum value of NPV was 0.9988. In comparison, the sensitivity and PPV values were relatively small, whose mean values are less than 0.4 (sensitivity: 0.3215, PPV: 0.3938). (2) For each scheme, the standard deviations of AUC and specificity were small (AUC: 0.0430; specificity: 0.0208), which means that the values of them are stable.

Table 4.

Statistics on the performance metrics.

Max Min Mean St. Dev.
AUC 0.7199 0.5310 0.6537 0.0430
Accuracy 0.8578 0.5967 0.7166 0.1016
Sensitivity 0.6783 0.1736 0.3215 0.1360
Specificity 0.9012 0.8419 0.8751 0.0208
PPV 0.6383 0.0031 0.3938 0.2362
NPV 0.9988 0.5914 0.7760 0.1634

Accuracy: ratio of the correctly predicted sample number to the total predicted sample number. Sensitivity: ratio of correctly predicted positive samples to the total number of true positive samples. Specificity: ratio of correctly predicted negative sample number to the total number of true negative samples. PPV: positive predictive value. NPV: negative predictive value. AUC: area under the receiver operating characteristic curve.

3.2. Analysis at Factor Level

AUC is a useful metric that comprehensively reflects the performance of the model. The average AUCs grouped by each factor (feature selection strategy, machine learning model, and sampling method) are shown in Table 5. The average AUCs grouped by feature selection strategy indicate the following: (1) schemes using LASSO-based strategy had the largest AUC mean (mean: 0.6582), and these using forward selection strategy had the smallest (mean: 0.6426). However, the difference is quite small (0.0156). (2) In terms of standard deviations, except for the forward selection strategy (standard deviation: 0.0626), other strategies are of little difference between each other, and all of them were less than 0.0450. (3) The range (gap between maximum and minimum) of average AUC value of each strategy was around 0.2220.

Table 5.

Statistics on AUCs grouped by different factors.

Grouping factors Groups N Mean St. Dev. Min Pctl(25) Pctl(75) Max Range
Feature selection strategy Original strategy 75 0.6547 0.0428 0.5125 0.6372 0.6819 0.7292 0.2167
LASSO-based strategy 75 0.6582 0.0441 0.4948 0.6449 0.6821 0.7239 0.2291
Forward selection strategy 75 0.6426 0.0626 0.5085 0.6307 0.6846 0.7355 0.2270
Backward selection strategy 75 0.6580 0.0427 0.5152 0.6435 0.6788 0.7331 0.2179
Importance-based strategy 75 0.6550 0.0436 0.5125 0.6350 0.6823 0.7307 0.2182

Machine learning model RF 75 0.6889 0.0314 0.6100 0.6713 0.7129 0.7355 0.1255
LR 75 0.6666 0.0222 0.6297 0.6514 0.6792 0.7161 0.0864
XGBoost-tree 75 0.6574 0.0256 0.6110 0.6353 0.6748 0.7171 0.1061
NNET 75 0.6552 0.0259 0.6163 0.6362 0.6702 0.7239 0.1076
SVM-linear 75 0.6005 0.0668 0.4948 0.5360 0.6558 0.7184 0.2236

Sampling method Oversampling 125 0.6604 0.0403 0.5085 0.6385 0.6908 0.7355 0.2270
Undersampling 125 0.6548 0.0354 0.5085 0.6398 0.6756 0.7239 0.2154
Original sampling 125 0.6460 0.0626 0.4948 0.6365 0.6911 0.7340 0.2392

N: number of cases. Mean: mean value corresponding to AUC of each model. St. Dev.: standard deviation corresponding to AUC of each model. Pctl(25): AUC corresponds to the first quartile of the variance numerical distribution of each model. Pctl(75): AUC corresponds to the third quartile of the variance numerical distribution of each model. LR: logistic regression model. NNET: neural networks. RF: random forest. SVM-linear: support vector machine-linear. XGBoost-tree: extreme gradient boosting-tree.

The average AUCs grouped by machine learning model indicate the following: (1) RF model had the largest AUC mean (mean: 0.6889). (2) The ascending standard deviations were 0.0222 (LR), 0.0256 (XGBoost-tree), 0.0259 (NNET), 0.0314 (RF), and 0.0668 (SVM-linear models). (3) Compared with the other models, the average AUC of the RF model had the largest maximum (max: 0.7355).

The average AUCs grouped by sampling method indicate the following: (1) for oversampling and original sampling methods, AUC had the largest and the least mean values, respectively (oversampling: 0.6604; original sampling: 0.6460). (2) Among the three sampling methods, undersampling achieved the least standard deviation (standard deviation: 0.0354).

For comprehensively evaluating the performance of the factors mentioned above, one-sided Delong tests were conducted on the ROCs of each paired performance group; and the superiority index, which refers to the percentage of results of a certain group which precedes that of another group, is used to indicate the superiority between each factor. The superiority indexes are summarized in Tables 68 in the aspect of feature selection strategy, machine learning models, and sampling method, respectively, and only the factor-comparison groups with more than 1/2 superiority index were analyzed accordingly.

Table 6.

Superiority index of Delong test on AUCs grouped by feature selection strategies.

Original strategy LASSO-based strategy Forward selection strategy Backward selection strategy Importance-based strategy
Original strategy NA 23/75 8/75 8/75 24/75
LASSO-based strategy 13/75 NA 10/75 11/75 14/75
Forward selection strategy 11/75 21/75 NA 13/75 11/75
Backward selection strategy 0/75 18/75 5/75 NA 10/75
Importance-based strategy 11/75 21/75 2/75 10/75 NA

NA: not available.

Table 7.

Superiority index of Delong test on AUCs grouped by machine learning model.

RF LR XGBoost-tree SVM-linear NNET
RF NA 7/75 4/75 3/75 5/75
LR 35/75 NA 2/75 0/75 1/75
XGBoost-tree 45/75 7/75 NA 1/75 2/75
SVM-linear 56/75 48/75 46/75 NA 46/75
NNET 49/75 12/75 5/75 7/75 NA

LR: logistic regression model. NNET: neural networks. RF: random forest. SVM-linear: support vector machine-linear. XGBoost-tree: extreme gradient boosting-tree. NA: not available.

Table 8.

Superiority index of Delong test on AUCs grouped by sampling method.

Oversampling Undersampling Original sampling
Oversampling NA 8/125 18/125
Undersampling 18/125 NA 25/125
Original sampling 20/125 18/125 NA

NA: not available.

According to Tables 6 and 8, feature selection methods and sampling methods show no factor-comparison group with a more than 1/2 superiority index, which indicates that there is no factor (such as oversampling) superior to another factor (such as undersampling) with a threshold of 1/2. Table 7 presents the superiority index on machine learning models, and it indicates the following: (1) Compared with other machine learning methods, the superiority indexes of RF compared with other machine learning methods are all more than 1/2 (RF versus XGBoost-tree: 45/75, RF versus SVM-linear-tree: 56/75, and RF versus NNET: 49/75), except for the LR; however, the superiority index of RF compared to LR is 35/75, which is very close to 1/2 and much higher than that of LR compared to RF. The analysis above indicates the superiority of RF. (2) All superiority indexes of other machine learning methods compared to SVM-linear are more than 1/2, which indicates the inferiority of SVM-linear.

Table 9 summarized the BGSS, WGSS, and TSS of AUCs grouped by three different factors mentioned above and indicated the following: (1) The maximum of BGSS (0.3191) and minimum of WGSS (0.5376) were obtained by machine learning models' grouping, which accounted for 37.25% and 62.75% of TSS, respectively. (2) The BGSS and WGSS grouped by sampling method were similar to the counterparts grouped by feature selection strategy (BGSS and WGSS grouped by sampling method: 0.0132 and 0.8435; BGSS and WGSS grouped by feature selection strategy: 0.0123 and 0.8443).

Table 9.

Analysis of sum of square error.

TSS BGSS BGSS/TSS (%) WGSS WGSS/TSS (%)
Grouped by feature selection strategy 0.8567 0.0123 1.44 0.8443 98.55
Grouped by ML model 0.8567 0.3191 37.25 0.5376 62.75
Grouped by sampling method 0.8567 0.0132 1.54 0.8435 98.46

TTS: total sum of square error. BGSS: between-groups sum of square error. WGSS: within-groups sum of square error.

4. Discussion

The present results indicate the feasibility and robustness of identifying elective urologic surgeries with high cancellation risk. The average AUCs in the test set exceeded 0.65, with the maximum of AUC (0.7199, RF, original sampling, and backward selection strategy).

It is worth noting that RF model accounted for 11 of the top 15 among the 75 schemes, and the schemes with a higher than 0.7 AUC were all RF models. In addition, RF model was significantly different from the other models in the aspect of AUC. Finally, machine learning model factor achieved the maximum of BGSS and the minimum of WGSS. Hence, the selection of machine learning models is considered a key factor in identification of surgeries with high cancellation risk.

Among the top 8 schemes, the first four and the last four were original sampling and oversampling, respectively. However, sampling methods show no significant difference with each other according to the Delong test. In addition, there were large WGSS and small BGSS when grouped by sampling methods. In summary, sampling method is the factor that affects the identification of surgeries with high cancellation risk, but not as important as the machine learning methods, to some extent.

For different feature selection strategies, differences between their means were little and insignificant. In addition, there were large WGSS and small BGSS when grouped by sampling methods. Hence, feature selection strategy is not an important factor that affects the identification of surgeries with high cancellation risk.

This study has limitations. It only focused on the elective urologic surgeries in one hospital, which means that the data are single-centered. Although our predictors covered five information categories that previous studies have covered and on which good results have been obtained, there are also potentially useful predictors that have not been collected. Further research can consider conducting multicenter studies, including multiple hospitals and departments.

5. Conclusion

This study provided effective methods for the identification of surgeries with high cancellation risk based on machine learning models and analyzed the key factors that affect the identification of surgeries with high cancellation risk. It proved the feasibility and robustness of identifying surgeries with high cancellation risk, with the considerable maximum of AUC (0.7199) for RF with original sampling using a backward selection strategy. In addition, two-sided test and sum of square error analysis were conducted to measure the effects of feature selection strategy, machine learning model, and sampling method on the identification of surgeries with high cancellation risk; and the selection of machine learning model was identified as the key factors that affect the identification of surgeries with high cancellation risk. This study offers methodology and insights for identifying the key experimental factors for identifying surgery cancellations, and the insights of it are useful to further research on machine learning-based identification of surgeries with high cancellation risk, in designing experimental process.

Acknowledgments

The work was supported in part by the National Natural Science Foundation of China (nos. 71532007, 71131006, and 71172197) and funded by the Sichuan University (2018skzx-pt173 and skzx2017-sb222) and the Innovation Project of Guangxi Graduate Education (YCSZ2014203).

Data Availability

The data supporting the study findings will not be shared since it is an organizational property. Data were anonymous, and study subjects could not be identified.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Authors' Contributions

Fengyi Zhang conceived and designed the experiments; Fengyi Zhang performed the experiments; Renrong Gong contributed to reagents/materials/analysis tools; Xinyuan Cui, Chuan Zhang, and Zhigao Liao analyzed the data; Fengyi Zhang, Chuan Zhang, and Xinyuan Cui drafted or revised the manuscript; Fengyi Zhang and Xinyuan Cui approved the final version.

References

  • 1.Hovlid E. A new pathway for elective surgery to reduce cancellation rates. BMC Health Services Research. 2012;12 doi: 10.1186/1472-6963-12-154. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Haana V., Sethuraman K., Stephens L., Rosen H., Meara J. G. Case cancellations on the day of surgery: an investigation in an Australian paediatric hospital. ANZ Journal of Surgery. 2009;79(9):636–640. doi: 10.1111/j.1445-2197.2009.05019.x. [DOI] [PubMed] [Google Scholar]
  • 3.Hovlid E. Patient experiences with interventions to reduce surgery cancellations: a qualitative study. BMC Surgery. 2013;13 doi: 10.1186/1471-2482-13-30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Luo L., Zhang F., Yao Y., Gong R., Fu M., Xiao J. Machine learning for identification of surgeries with high risks of cancellation. Health Informatics Journal. 2018;26(1):141–155. doi: 10.1177/1460458218813602. [DOI] [PubMed] [Google Scholar]
  • 5.Macario A., Dexter F., Traub R. D. Hospital profitability per hour of operating room time can vary among surgeons. Anesthesia & Analgesia. 2001;93(3):669–675. doi: 10.1097/00000539-200109000-00028. [DOI] [PubMed] [Google Scholar]
  • 6.Dexter F., Blake J. T., Penning D. H., Lubarsky D. A. Calculating a potential increase in hospital margin for elective surgery by changing operating room time allocations or increasing nursing staffing to permit completion of more cases: a case study. Anesthesia & Analgesia. 2002;94(1):138–142. doi: 10.1213/00000539-200201000-00026. [DOI] [PubMed] [Google Scholar]
  • 7.Lavery A. M., Banwell B. L., Liu G., Waldman A. T. Hospital admission rates for pediatric multiple sclerosis in the United States using the pediatric health information system (PHIS) Multiple Sclerosis and Related Disorders. 2016;9:5–10. doi: 10.1016/j.msard.2016.05.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Olson D., Todd J. K., Birkholz M., Asturias E. J., Gaensbauer J. T. Analysis of the pediatric health information system database as a surveillance tool for travel-associated infectious diseases. The American Journal of Tropical Medicine and Hygiene. 2015;92(5):1067–1069. doi: 10.4269/ajtmh.14-0794. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Jia C.-H. Analysis of clinical medication rules in 48 398 patients with limb fractures based on hospital information system. China Journal of Chinese Materia Medica. 2016;41(13):2555–2560. doi: 10.4268/cjcmm20161331. [DOI] [PubMed] [Google Scholar]
  • 10.Atramont A., Bonnet-Zamponi D., Bourdel-Marchasson I., Tangre I., Fagot-Campagna A., Tuppin P. Health status and drug use 1 year before and 1 year after skilled nursing home admission during the first quarter of 2013 in France: a study based on the French national health insurance information system. European Journal of Clinical Pharmacology. 2018;74(1):109–118. doi: 10.1007/s00228-017-2343-y. [DOI] [PubMed] [Google Scholar]
  • 11.Kim E. H. Application and evaluation of personal health information management system. Proceedings of the 26th Annual International Conference of the Ieee Engineering in Medicine and Biology Society; September 2004; San Francisco, CA, USA. pp. 3159–3162. [DOI] [PubMed] [Google Scholar]
  • 12.Wang M., Lau C., Matsen F. A., Kim Y. Personal health information management system and its application in referral management. IEEE Transactions on Information Technology in Biomedicine. 2004;8(3):287–297. doi: 10.1109/titb.2004.834397. [DOI] [PubMed] [Google Scholar]
  • 13.Bittencourt S. A., Camacho L. A. B., Leal M. d. C. O Sistema de Informação Hospitalar e sua aplicação na saúde coletiva. Cadernos de Saúde Pública. 2006;22(1):19–30. doi: 10.1590/s0102-311x2006000100003. [DOI] [PubMed] [Google Scholar]
  • 14.Li H., Wu J., Gao Y., Shi Y. Examining individuals’ adoption of healthcare wearable devices: an empirical study from privacy calculus perspective. International Journal of Medical Informatics. 2016;88:8–17. doi: 10.1016/j.ijmedinf.2015.12.010. [DOI] [PubMed] [Google Scholar]
  • 15.Eden K. B., Totten A. M., Kassakian S. Z., et al. Barriers and facilitators to exchanging health information: a systematic review. International Journal of Medical Informatics. 2016;88:44–51. doi: 10.1016/j.ijmedinf.2016.01.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Wu X., Wu C., Zhang K., Wei D. Residents’ numeric inputting error in computerized physician order entry prescription. International Journal of Medical Informatics. 2016;88:25–33. doi: 10.1016/j.ijmedinf.2016.01.002. [DOI] [PubMed] [Google Scholar]
  • 17.González-Arévalo A., Gómez-Arnau J. I., delaCruz F. J., et al. Causes for cancellation of elective surgical procedures in a Spanish general hospital. Anaesthesia. 2009;64(5):487–493. doi: 10.1111/j.1365-2044.2008.05852.x. [DOI] [PubMed] [Google Scholar]
  • 18.Schofield W. N., Rubin G. L., Piza M., et al. Cancellation of operations on the day of intended surgery at a major Australian referral hospital. Medical Journal of Australia. 2005;182(12):612–615. doi: 10.5694/j.1326-5377.2005.tb06846.x. [DOI] [PubMed] [Google Scholar]
  • 19.Al Talalwah N., McIltrot K. H. Cancellation of surgeries: integrative review. Journal of Perianesthesia Nursing. 2019;34(1):86–96. doi: 10.1016/j.jopan.2017.09.012. [DOI] [PubMed] [Google Scholar]
  • 20.Sultan N., Rashid A., Abbas S. M. Reasons for cancellation of elective cardiac surgery at prince sultan cardiac centre, Saudi Arabia. Journal of the Saudi Heart Association. 2012;24(1):29–34. doi: 10.1016/j.jsha.2011.10.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Cho H. S., Lee Y. S., Lee S. G., Kim J. M., Kim T. H. Reasons for surgery cancellation in a general hospital: a 10-year study. International Journal of Environmental Research and Public Health. 2019;16(1) doi: 10.3390/ijerph16010007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Yu K. Y. Contributing factors of elective surgical case cancellation: a retrospective cross-sectional study at a single-site hospital. BMC Surgery. 2017;17 doi: 10.1186/s12893-017-0296-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Yusuff U., Handa V. Reasons for day of surgery cancellations: identifying trends to improve efficiency. Anaesthesia. 2017;72:p. 81. [Google Scholar]
  • 24.Seim A. R., Fagerhaug T., Ryen S. M., et al. Causes of cancellations on the day of surgery at two major university hospitals. Surgical Innovation. 2009;16(2):173–180. doi: 10.1177/1553350609335035. [DOI] [PubMed] [Google Scholar]
  • 25.Kaddoum R. Causes of cancellations on the day of surgery at a tertiary teaching hospital. BMC Health Services Research. 2016;16 doi: 10.1186/s12913-016-1475-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Turunen E., Miettinen M., Setälä L., Vehviläinen-Julkunen K. Elective surgery cancellations during the time between scheduling and operation. Journal of Perianesthesia Nursing. 2019;34(1):97–107. doi: 10.1016/j.jopan.2017.09.014. [DOI] [PubMed] [Google Scholar]
  • 27.Azari-Rad S., Yontef A., Aleman D., Urbach D. Reducing elective general surgery cancellations at a Canadian hospital. Canadian Journal of Surgery. 2013;56(2):113–118. doi: 10.1503/cjs.018411. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Morrice D. A simulation analysis of a patient-centered surgical home to improve outpatient surgical processes of care and outcomes. Proceedings of the 2013 Winter Simulation Conference; December 2013; Washington, DC, USA. pp. 2274–2286. [Google Scholar]
  • 29.Tan A. L., Chiew C. J., Wang S., et al. Risk factors and reasons for cancellation within 24 h of scheduled elective surgery in an academic medical centre: a cohort study. International Journal of Surgery. 2019;66:72–78. doi: 10.1016/j.ijsu.2019.04.009. [DOI] [PubMed] [Google Scholar]
  • 30.Wong D. J. N., Harris S. K., Moonesinghe S. R. Cancelled operations: a 7-day cohort study of planned adult inpatient surgery in 245 UK national health service hospitals. British Journal of Anaesthesia. 2018;121(4):730–738. doi: 10.1016/j.bja.2018.07.002. [DOI] [PubMed] [Google Scholar]
  • 31.Daghistani T. A., Elshawi R., Sakr S., Ahmed A. M., Al-Thwayee A., Al-Mallah M. H. Predictors of in-hospital length of stay among cardiac patients: a machine learning approach. International Journal of Cardiology. 2019;288:140–147. doi: 10.1016/j.ijcard.2019.01.046. [DOI] [PubMed] [Google Scholar]
  • 32.Nithya B. V. Predictive analytics in health care using machine learning tools and techniques. Proceedings of the 2017 International Conference on Intelligent Computing and Control Systems; June 2017; Madurai, India. pp. 492–499. [Google Scholar]
  • 33.Kocsis O. Assessing machine learning algorithms for self-management of asthma. Proceedings of the 2017 IEEE International Conference on E-Health and Bioengineering Conference; June 2017; Sinaia, Romania. pp. 571–574. [Google Scholar]
  • 34.Bertsimas D., Bjarnadóttir M. V., Kane M. A., et al. Algorithmic prediction of health-care costs. Operations Research. 2008;56(6):1382–1392. doi: 10.1287/opre.1080.0619. [DOI] [Google Scholar]
  • 35.Karnuta J. M., Navarro S. M., Haeberle H. S., et al. Predicting inpatient payments prior to lower extremity arthroplasty using deep learning: which model architecture is best? The Journal of Arthroplasty. 2019;34(10):2235–2241. doi: 10.1016/j.arth.2019.05.048. [DOI] [PubMed] [Google Scholar]
  • 36.Kose I., Gokturk M., Kilic K. An interactive machine-learning-based electronic fraud and abuse detection system in healthcare insurance. Applied Soft Computing. 2015;36:283–299. doi: 10.1016/j.asoc.2015.07.018. [DOI] [Google Scholar]
  • 37.Seo J. O. Identifying frauds and anomalies in medicare-B dataset. Proceedings of the 2017 39th Annual International Conference of the Ieee Engineering in Medicine and Biology Society; July 2017; Jeju Island, Republic of Korea. pp. 3664–3667. [DOI] [PubMed] [Google Scholar]
  • 38.Chimmad A., Saripalli P., Tirumala V. Assessment of healthcare claims rejection risk using machine learning. Proceedings of the 2017 IEEE 19th International Conference on E-Health Networking, Applications and Services (Healthcom); October 2017; Dalian, China. pp. 1–6. [Google Scholar]
  • 39.Swangnetr M., Kaber D. B. Emotional state classification in patient-robot interaction using wavelet analysis and statistics-based feature selection. Ieee Transactions on Human-Machine Systems. 2013;43(1):63–75. doi: 10.1109/tsmca.2012.2210408. [DOI] [Google Scholar]
  • 40.Kim K. H., Ku B., Kang N., Kim Y. S., Jang J. S., Kim J. Y. Study of a vocal feature selection method and vocal properties for discriminating four constitution types. Evidence-Based Complementary and Alternative Medicine. 2012;2012:10. doi: 10.1155/2012/831543.831543 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Liu L., Ni Y., Zhang N., Pratap J. Mining patient-specific and contextual data with machine learning technologies to predict cancellation of children’s surgery. International Journal of Medical Informatics. 2019;129:234–241. doi: 10.1016/j.ijmedinf.2019.06.007. [DOI] [PubMed] [Google Scholar]
  • 42.Fathima A. S., Manimeglai D. Analysis of significant factors for dengue infection prognosis using the random forest classifier. International Journal of Advanced Computer Science and Applications. 2015;6(2):240–245. [Google Scholar]
  • 43.Byeon H. Developing a random forest classifier for predicting the depression and managing the health of caregivers supporting patients with Alzheimer’s Disease. Technology and Health Care. 2019;27(5):531–544. doi: 10.3233/thc-191738. [DOI] [PubMed] [Google Scholar]
  • 44.Bellini V. Artificial intelligence: a new tool in operating room management. Role of machine learning models in operating room optimization. Journal of Medical Systems. 2019;44(1) doi: 10.1007/s10916-019-1512-1. [DOI] [PubMed] [Google Scholar]
  • 45.ShahabiKargar Z. Predicting procedure duration to improve scheduling of elective surgery. In: Pham D. N., Park S. B., editors. Pricai 2014: Trends in Artificial Intelligence. Berlin, Germany: Springer; 2014. pp. 998–1009. [Google Scholar]
  • 46.Jaimes F., Farbiarz J., Alvarez D., Martínez C. Comparison between logistic regression and neural networks to predict death in patients with suspected sepsis in the emergency room. Critical Care. 2005;9(2):R150–R156. doi: 10.1186/cc3054. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Clermont G., Angus D. C., DiRusso S. M., Griffin M., Linde-Zwirble W. T. Predicting hospital mortality for patients in the intensive care unit: a comparison of artificial neural networks with logistic regression models. Critical Care Medicine. 2001;29(2):291–296. doi: 10.1097/00003246-200102000-00012. [DOI] [PubMed] [Google Scholar]
  • 48.Meyfroidt G., Güiza F., Ramon J., Bruynooghe M. Machine learning techniques to examine large patient databases. Best Practice & Research Clinical Anaesthesiology. 2009;23(1):127–143. doi: 10.1016/j.bpa.2008.09.003. [DOI] [PubMed] [Google Scholar]
  • 49.He Z. M. Quantum-enhanced feature selection with forward selection and backward elimination. Quantum Information Processing. 2018;17(7) doi: 10.1007/s11128-018-1924-8. [DOI] [Google Scholar]
  • 50.Marelli A. Machine learning developed algorithms to identify patients with congenital heart disease from large claims databases. Circulation. 2019;140(1) [Google Scholar]
  • 51.Huang N. T. Feature selection of power quality disturbance signals with an entropy-importance-based random forest. Entropy. 2016;18(2) doi: 10.3390/e18020044. [DOI] [Google Scholar]
  • 52.Wang H. Using random forests to select optimal input variables for short-term wind speed forecasting models. Energies. 2017;10(10) doi: 10.3390/en10101522. [DOI] [Google Scholar]
  • 53.Shi D. H., Guan J., Zurada J. Cost-sensitive learning for imbalanced bad debt datasets in healthcare industry. In: Monsalve C., Sanchez A., editors. Proceedings of the 2015 Asia-Pacific Conference on Computer-Aided System Engineering—Apcase 2015; July 2015; Quito, Ecuador. pp. 30–35. [Google Scholar]
  • 54.Naveen N., Ravi V., Kumar D. A. Application of fuzzyARTMAP for churn prediction in bank credit cards. International Journal of Information and Decision Sciences. 2009;1(4):428–444. doi: 10.1504/ijids.2009.027761. [DOI] [Google Scholar]
  • 55.Santoso B. A comparative study of synthetic over-sampling method to improve the classification of poor households in yogyakarta province. IOP Conference Series: Earth and Environmental Science. 2018;187 doi: 10.1088/1755-1315/187/1/012048.012048 [DOI] [Google Scholar]
  • 56.Feng L. Analysis of the influencing factors of elective surgery cancellation. In: Li S., Dai Y., Cheng Y., editors. Proceedings of the 2015 7th International Conference on Information Technology in Medicine and Education; November 2015; Huangshan, China. pp. 41–45. [Google Scholar]
  • 57.Zalawadiya S. K., Gunasekaran P. C., Bavishi C. P., Veeranna V., Panaich S., Afonso L. Left ventricular hypertrophy and risk reclassification for coronary events in multi-ethnic adults. European Journal of Preventive Cardiology. 2015;22(5):673–679. doi: 10.1177/2047487314530383. [DOI] [PubMed] [Google Scholar]
  • 58.Thomas L. E., O’Brien E. C., Piccini J. P., D’Agostino R. B., Pencina M. J. Application of net reclassification index to non-nested and point-based risk prediction models: a review. European Heart Journal. 2019;40(23):1880–1887. doi: 10.1093/eurheartj/ehy345. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Jayasekara D. N., Sooriyarachchi M. R. A simulation based study for comparing tests associated with receiver operating characteristic (ROC) curves. Communications in Statistics—Simulation and Computation. 2014;43(10):2444–2467. doi: 10.1080/03610918.2012.752840. [DOI] [Google Scholar]
  • 60.Chen X. Prediction of optimal gene functions for osteosarcoma using gene ontology and microarray profiles. Journal of Bone Oncology. 2017;7:18–22. doi: 10.1016/j.jbo.2017.04.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Yang J. Differential diagnosis of pancreatic serous cystadenoma and mucinous cystadenoma: utility of textural features in combination with morphological characteristics. BMC Cancer. 2019;19(1):p. 1223. doi: 10.1186/s12885-019-6421-7. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The data supporting the study findings will not be shared since it is an organizational property. Data were anonymous, and study subjects could not be identified.


Articles from Journal of Healthcare Engineering are provided here courtesy of Wiley

RESOURCES