Key Points
Question
Can machine learning models accurately predict patient disability and pain following lumbar disc herniation surgery?
Findings
In this prognostic study including 22 707 patients, machine learning models were developed and validated in large-scale, nationally representative data for treatment success or nonsuccess in disability and pain 12 months after lumbar disc herniation surgery. The models showed good discrimination and calibration.
Meaning
The findings of this study suggest that algorithms can inform about individual prognosis and aid in surgical decision-making to ultimately reduce ineffective and costly spine care.
This prognostic study develops machine learning tools to use for prediction of patient pain and disability following lumbar disc herniation surgery.
Abstract
Importance
Lumber disc herniation surgery can reduce pain and disability. However, a sizable minority of individuals experience minimal benefit, necessitating the development of accurate prediction models.
Objective
To develop and validate prediction models for disability and pain 12 months after lumbar disc herniation surgery.
Design, Setting, and Participants
A prospective, multicenter, registry-based prognostic study was conducted on a cohort of individuals undergoing lumbar disc herniation surgery from January 1, 2007, to May 31, 2021. Patients in the Norwegian Registry for Spine Surgery from all public and private hospitals in Norway performing spine surgery were included. Data analysis was performed from January to June 2023.
Exposures
Microdiscectomy or open discectomy.
Main Outcomes and Measures
Treatment success at 12 months, defined as improvement in Oswestry Disability Index (ODI) of 22 points or more; Numeric Rating Scale (NRS) back pain improvement of 2 or more points, and NRS leg pain improvement of 4 or more points. Machine learning models were trained for model development and internal-external cross-validation applied over geographic regions to validate the models. Model performance was assessed through discrimination (C statistic) and calibration (slope and intercept).
Results
Analysis included 22 707 surgical cases (21 161 patients) (ODI model) (mean [SD] age, 47.0 [14.0] years; 12 952 [57.0%] males). Treatment nonsuccess was experienced by 33% (ODI), 27% (NRS back pain), and 31% (NRS leg pain) of the patients. In internal-external cross-validation, the selected machine learning models showed consistent discrimination and calibration across all 5 regions. The C statistic ranged from 0.81 to 0.84 (pooled random-effects meta-analysis estimate, 0.82; 95% CI, 0.81-0.84) for the ODI model. Calibration slopes (point estimates, 0.94-1.03; pooled estimate, 0.99; 95% CI, 0.93-1.06) and calibration intercepts (point estimates, −0.05 to 0.11; pooled estimate, 0.01; 95% CI, −0.07 to 0.10) were also consistent across regions. For NRS back pain, the C statistic ranged from 0.75 to 0.80 (pooled estimate, 0.77; 95% CI, 0.75-0.79); for NRS leg pain, the C statistic ranged from 0.74 to 0.77 (pooled estimate, 0.75; 95% CI, 0.74-0.76). Only minor heterogeneity was found in calibration slopes and intercepts.
Conclusion
The findings of this study suggest that the models developed can inform patients and clinicians about individual prognosis and aid in surgical decision-making.
Introduction
The volume of lumbar spine surgery has increased considerably over the past decades, placing large costs on health care systems.1,2,3 Lumbar disc herniation surgery effectively reduces disability and pain in most patients, but a subset experience minimal benefit.4,5,6 In Scandinavia, 24% to 32% of patients do not achieve an important improvement in pain-related disability 1 year postoperatively.7 In most cases, the indication for surgery is relative. Therefore, shared decision-making, appraising both potential treatment risks and benefits, is essential to minimize ineffective and costly spine care.8,9 Providing precise probabilities of outcomes based on individual patient characteristics in a presurgical setting would allow clinicians to manage patients’ expectations before surgery and help patients make an informed choice about surgery.
Prediction models for disability and pain improvements following degenerative spine surgery have been developed; however, most studies including patients with lumbar disc herniation have limited generalizability due to a low number of patients from single surgical centers.10,11,12,13,14,15 Population-based spine registries with near complete national coverage hold a unique potential for prognostic modeling due to their comprehensive inclusion of a broad range of presurgical variables. Moreover, they reflect clinical practice settings and account for the uniqueness of a specific patient population.16,17
The volume and complexity of data available in national spine registries provide opportunities to develop better prediction models, which is necessary to improve quality of spine care.16 Machine learning algorithms are powerful tools for analyzing large amounts of data and have gained traction in recent years, but their use for outcome prediction in spine surgery remains nascent.18 The hope is that machine learning based on large and representative data can predict outcomes with high accuracy and consequently assist clinicians and patients in weighing the risks and benefits of surgical intervention. Therefore, the purpose of this study was to develop and validate machine learning models for predicting improvement in disability and pain 12 months after lumbar disc herniation surgery. We used internal-external cross-validation to evaluate generalizability over 4 geographic regions in Norway and a separate cluster for private hospitals.
Methods
Design
This was a multicenter study using prospectively collected data from adults undergoing surgery for lumbar disc herniation included in the Norwegian Registry for Spine Surgery (NORspine). We followed the methodologic framework proposed by the Prognosis Research Strategy group19 and report the study in line with the Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) reporting guideline.20 This study is part of the AID-Spine project, which has been approved by the ethics committee of the Health Region of South-East Norway. Written informed consent was obtained from all patients in the NORspine, and the Data Protection Authority of Norway approved the registry protocol.
Data Source and Patient Population
NORspine is a comprehensive clinical registry for degenerative spine surgery designed for quality control and research. The register is mandatory and had a coverage of 100% at the surgical unity level in 2021 (40 centers). The individual-level completeness rate was 81% for lumbar spine surgery in 2021.21 All patients included in the NORspine registry who had undergone lumbar discectomy from January 1, 2007, to May 31, 2021, were identified and screened for eligibility. Repeat surgeries were included as new cases if performed more than 90 days after the index surgery. Reoperation within 90 days is considered a complication in the NORspine registry, and thus not recorded as a new case. The registry also excludes patients who undergo an operation for fractures, trauma, or cancer.21 Patients with cauda equina syndrome were excluded from the current study.
The NORspine data collection process comprised a preoperative form including patient-reported outcomes to be completed by the patients at the time of surgical admission (baseline). Information regarding diagnosis, previous lumbar spine surgery, comorbidity, imaging findings, and surgical procedure were recorded by the surgeon on a standardized form. At 12 months postoperatively, follow-up questionnaires including patient-reported outcomes were distributed by mail to the patients.
Outcomes
The outcomes included measured patient improvements using the Oswestry Disability Index (ODI)22 and Numeric Rating Scale (NRS)23 for back pain and leg pain from baseline to 12 months. The ODI was the primary outcome; it is a 10-item score from 0 (none) to 100 (maximum disability) encompassing limitations in various activities of daily living.22 The NRS measures pain intensity during the last week on an 11-point scale, with 0 representing no pain and 10 the worst imaginable pain.23
The outcomes were operationalized as treatment success, with study-specific calculations of the cutoffs for success. The thresholds were arrived at using the anchor-based predictive modeling method,24 adjusted for the proportion of patients reporting improvement.25 As the dichotomized anchor response, we used a 7-point Global Perceived Effect scale,26 with the cutoff for success and nonsuccess set between patients responding they were much improved vs slightly improved. The cutoff scores were: ODI, 22 points; NRS back pain, 2 points; and NRS leg pain, 4 points improvement from baseline to 12 months.
Predictors
We used all potential preoperative predictors available in the NORspine registry (eTable 1 in Supplement 1 provides a detailed description). In short, 25 features (predictor variables) were included, covering patient demographic characteristics, comorbidity, clinical characteristics, analgesics use, and type of operation. Only preoperative features were included, given our aim of improving the selection of surgical candidates.
Sample Size
Our sample size was restricted to available NORspine data. An a priori sample size calculation was performed to evaluate the adequacy of the data set and guide how many predictors could be included.27 We assumed an event rate of 30% (nonsuccess at 12 months),7 a C statistic of 0.75 based on a recent systematic review,28 and a maximum number of 50 predictor parameters. Based on these inputs, a sample size of 2551 patients would be required for the model development, corresponding to 766 events and an event per parameter of 15.3. The pmsampsize package in Stata, version 17.1 (StataCorp LLC) was used for the calculations,27 with Cox-Snell R2 value estimated from the C statistic.29 While the number of predictor parameters for machine learning likely exceeds that for regression, our sample size far exceeds the minimum requirement estimate for regression-based prediction models.27
Data Cleaning and Quality Checks
The NORspine registry data quality is periodically assessed by the registry owner to detect systematic or random errors in the data entry.30 We performed further data quality checks during data cleaning, including assessment of potential duplicate entries, outliers, the extent of missing data, and presence of systematically missing variables within and across clusters. All patient characteristics and model predictors were determined at the time of surgical admission. The same eligibility criteria and characteristic determination methods were applied to all clusters.
Statistical Analysis
Data analysis was performed from January to June 2023. We calculated descriptive statistics for baseline characteristics, overall and for each cluster separately. We applied multiple imputation with chained equations to handle missing baseline and outcome data, which were assumed to be missing at random, with 50 imputed data sets generated. The imputation models included all features and outcomes, performed separately for each cluster to allow the distribution of the imputed values to differ among clusters.31 Imputations were assessed for consistency by comparing distributions of imputed values with the complete data. The predictive performance measures were estimated in each imputed data set separately before being combined across imputations using the Rubin rule.32
Seven supervised machine learning algorithms were trained to develop the prediction model: random forest, logistic regression, linear discriminant analysis, multilayer perceptron, gradient boosting, extra trees, and extreme gradient boosting. Preprocessing steps involved scaling of continuous variables (minimum-maximum normalization) and 1-hot encoding of categorical variables. All features were included, ie, no variable selection techniques were used. Continuous variables were kept continuous to avoid loss of prognostic information. Hyperparameters were tuned using a grid search with 5-fold cross-validation (eTable 2 in Supplement 1). The best algorithm for each outcome was estimated based on model discrimination using the C statistic. We assessed apparent performance (using the same data in which the model was developed), quantified with the C statistic.
Internal-external cross-validation was used to evaluate the derived prediction models to give a more realistic estimate of model performance and heterogeneity in performance across regions.33,34 Internal-external cross-validation involves a nonrandom split of data based on clusters, in our case, 4 geographic regions corresponding to the 4 Norwegian Regional Health Authorities and an additional cluster for private hospitals. A single internal-external cross-validation cycle separates the data into a development cohort and validation cohort, with 4 of 5 clusters forming the development cohort and reserving the other cluster for validation. The process was repeated 5 times, each time reserving a different cluster for validation.
We calculated C statistics, positive predictive value, negative predictive value, calibration slopes, and calibration intercepts in each cluster. Overall performance measures with 95% CIs (derived using the Hartung-Knapp-Sidik-Jonkman variance correction) and 95% prediction intervals were also summarized across clusters using a random-effects meta-analysis.35,36 We further present calibration plots with comparison of observed to predicted risk for each model overall and by validation cluster (created using pmcalplot on Stata), generated separately in each imputed data set and checked for consistency across imputations. Clinical utility was examined using decision curve analysis by comparing the prediction models against blanket treatment strategies to treat all or to treat none at varying risk thresholds.37 Shapley Additive Explanations values were calculated for all 3 prediction models to investigate feature importance.38 The machine learning algorithms were implemented using Python 3.8.13 and Scikit-learn Python libraries. Stata was used for data cleaning and multiple imputation. As a sensitivity analysis, results obtained from using imputed data were compared with those of complete case analysis for each outcome.
Results
Of 56 963 surgical cases screened, we identified 22 707 surgical cases (21 161 patients) (mean [SD] age, 47.0 [14.0] years; 12 952 [57.0%] males; 9755 females [43.0%]) who underwent operations for lumbar disc herniation eligible for inclusion in our primary analysis (Figure 1). Baseline characteristics of the total study population and stratified by cluster for the ODI model are summarized in the Table. The analysis for NRS back pain included 23 804 cases and, for NRS leg pain, 22 691 cases. The proportions of cases experiencing treatment nonsuccess were 33% (ODI), 27% (NRS back pain), and 31% (NRS leg pain).
Figure 1. Flow Diagram of Surgical Cases Included in the Analysis.
NRS indicates Numeric Rating Scale; ODI, Oswestry Disability Index.
Table. Study Population for the ODI Model.
Characteristic | No. (%) | |||||
---|---|---|---|---|---|---|
Development, total (n = 22 707) | Cluster | |||||
Southern and Eastern (n = 6058) | Western (n = 4788) | Central (n = 4311) | Northern (n = 2138) | Private hospitals (n = 5412) | ||
ODIa | ||||||
Treatment success | 9934 (66.9) | 2363 (60.4) | 2032 (65.8) | 1966 (67.3) | 1086 (72.6) | 2487 (72.3) |
Missing | 7848 (34.6) | 2145 (35.4) | 1699 (35.5) | 1388 (32.2) | 643 (30.1) | 1973 (36.5) |
Sex | ||||||
Male | 12 952 (57.0) | 3139 (51.8) | 2540 (53.1) | 2339 (54.3) | 1247 (58.3) | 3687 (68.1) |
Female | 9755 (43.0) | 2919 (48.2) | 2248 (46.9) | 1972 (45.7) | 891 (41.7) | 1725 (31.9) |
Age, mean (SD), y | 47.0 (14.0) | 47.3 (14.3) | 47.9 (14.9) | 48.3 (14.7) | 45.3 (13.2) | 45.6 (12.2) |
BMI, mean (SD) | 27.0 (4.5) | 27.2 (4.7) | 26.9 (4.5) | 27.2 (4.5) | 27.3 (4.7) | 26.7 (4.0) |
Missing | 1327 (5.8) | 253 (4.2) | 113 (2.4) | 500 (11.6) | 84 (3.9) | 377 (7.0) |
Nonnative language speaker | 1617 (7.2) | 541 (9.0) | 393 (8.3) | 221 (5.2) | 120 (5.6) | 342 (6.4) |
Missing | 129 (0.6) | 45 (0.7) | 29 (0.6) | 28 (0.7) | 4 (0.2) | 23 (0.4) |
Marital status, single | 5583 (24.8) | 1670 (27.9) | 1216 (25.6) | 1017 (23.9) | 539 (25.4) | 1141 (21.2) |
Missing | 205 (0.9) | 71 (1.2) | 33 (0.7) | 55 (1.3) | 17 (0.8) | 29 (0.5) |
Smoker | 5797 (25.8) | 1604 (26.8) | 1273 (26.8) | 1134 (26.6) | 606 (28.7) | 1180 (22.0) |
Missing | 216 (1.0) | 71 (1.2) | 33 (0.7) | 55 (1.3) | 17 (0.8) | 29 (0.5) |
Education | ||||||
Lower secondary school | 3255 (14.5) | 902 (15.1) | 790 (16.7) | 692 (16.3) | 375 (17.7) | 496 (9.2) |
Upper secondary school | 10 575 (47.1) | 2793 (46.7) | 2237 (49.3) | 2072 (48.8) | 991 (46.8) | 2392 (44.5) |
University (1-3 y) | 4606 (20.5) | 1192 (20.0) | 855 (18.1) | 791 (18.6) | 385 (18.2) | 1383 (25.7) |
University (≥4 y) | 4003 (17.8) | 1089 (18.2) | 747 (15.8) | 692 (16.3) | 368 (17.4) | 1107 (20.6) |
Missing | 268 (1.2) | 82 (1.4) | 69 (1.4) | 64 (1.5) | 19 (0.9) | 34 (0.6) |
Work status | ||||||
Working/student | 5943 (26.8) | 1498 (25.4) | 1168 (24.8) | 972 (23.2) | 514 (24.7) | 1791 (34.0) |
Retirement ageb | 2277 (10.3) | 676 (11.5) | 618 (13.1) | 545 (13.0) | 156 (7.5) | 282 (5.4) |
Sick leave | 10 871 (49.1) | 2717 (46.1) | 2165 (45.9) | 1972 (47.1) | 1103 (53.0) | 2914 (55.4) |
Disability pension | 3058 (13.8) | 1005 (17.1) | 767 (16.3) | 702 (16.8) | 309 (14.8) | 275 (5.2) |
Missing | 558 (2.5) | 162 (2.7) | 70 (1.5) | 120 (2.8) | 56 (2.6) | 150 (2.8) |
Litigation issuec | 1723 (7.6) | 508 (8.4) | 320 (6.7) | 324 (7.5) | 168 (7.9) | 403 (7.5) |
Anxiety or depressiond | 8844 (39.8) | 2444 (41.4) | 1942 (41.3) | 1801 (42.7) | 873 (41.5) | 1784 (33.7) |
Missing | 490 (2.2) | 158 (2.6) | 90 (1.9) | 91 (2.1) | 34 (1.6) | 117 (2.2) |
Comorbidities | ||||||
0 | 16 922 (74.5) | 4366 (72.1) | 3277 (68.4) | 3098 (71.9) | 1709 (79.9) | 4472 (82.6) |
1 | 3879 (17.1) | 1153 (19.0) | 922 (19.3) | 805 (18.7) | 296 (13.8) | 703 (13.0) |
2 | 1363 (6.0) | 379 (6.3) | 406 (8.5) | 291 (6.8) | 100 (4.7) | 187 (3.5) |
≥3 | 543 (2.4) | 160 (2.6) | 183 (3.8) | 117 (2.7) | 33 (1.5) | 50 (0.9) |
ASA grade ≥3 | 1489 (6.6) | 456 (7.6) | 359 (7.6) | 441 (10.4) | 90 (4.4) | 143 (2.7) |
Missing | 297 (1.3) | 49 (0.8) | 34 (0.7) | 82 (1.9) | 71 (3.3) | 61 (1.1) |
ODI (0-100), mean (SD) | 48.6 (17.2) | 48.2 (16.8) | 50.9 (17.5) | 50.3 (17.8) | 52.5 (18.7) | 44.3 (15.2) |
Missing | 178 (0.8) | 57 (0.9) | 34 (0.7) | 23 (0.5) | 8 (0.4) | 56 (1.0) |
NRS pain intensitye | ||||||
Back pain | 6.6 (2.4) | 6.5 (2.30 | 6.8 (2.3) | 6.7 (2.3) | 6.9 (2.4) | 6.1 (2.3) |
Missing | 682 (3.0) | 217 (3.6) | 156 (3.3) | 120 (2.8) | 46 (2.2) | 143 (2.6) |
Leg pain | 7.2 (2.1) | 7.1 (2.1) | 7.4 (2.0) | 7.3 (2.1) | 7.5 (2.1) | 6.9 (2.0) |
Missing | 666 (2.9) | 206 (3.4) | 157 (3.3) | 130 (3.0) | 40 (1.9) | 133 (2.5) |
EQ-5D, mean (SD) | 0.48 (0.22) | 0.49 (0.22 | 0.45 (0.22) | 0.47 (0.22) | 0.44 (0.23) | 0.53 (0.19) |
Missing | 950 (4.2) | 307 (5.1) | 191 (4.0) | 197 (4.6) | 62 (2.9) | 193 (3.6) |
EQ VAS (0-100), mean (SD) | 43.2 (21.1) | 44.1 (21.1) | 40.6 (21.0) | 42.2 (21.5) | 41.8 (22.1) | 45.9 (20.1) |
Missing | 1278 (5.6) | 396 (6.5) | 306 (6.4) | 285 (6.6) | 83 (3.9) | 208 (3.8) |
Back pain, mo | ||||||
<3 | 4196 (19.2) | 765 (13.1) | 928 (19.9) | 848 (20.7) | 608 (29.3) | 1047 (20.2) |
3-11 | 9797 (44.8) | 2625 (45.1) | 2046 (43.8) | 1730 (42.1) | 800 (38.5) | 2596 (50.0) |
12-24 | 3166 (14.5) | 996 (17.1) | 678 (14.5) | 612 (14.9) | 271 (13.0) | 609 (11.7) |
>24 | 4715 (21.6) | 1441 (24.7) | 1015 (21.8) | 917 (22.3) | 399 (19.2) | 943 (18.2) |
Missing | 833 (3.7) | 231 (3.8) | 121 (2.5) | 204 (4.7) | 60 (2.8) | 217 (4.0) |
Leg pain, mo | ||||||
<3 | 5547 (25.6) | 1055 (18.3) | 1258 (27.3) | 1101 (27.1) | 773 (37.6) | 1360 (26.3) |
3-11 | 10 813 (49.9) | 2962 (51.5) | 2232 (48.4) | 1915 (47.1) | 847 (41.2) | 2857 (55.2) |
12-24 | 2760 (12.7) | 917 (15.9) | 597 (12.9) | 524 (12.9) | 228 (11.1) | 494 (9.5) |
>24 | 2548 (11.8) | 823 (14.3) | 527 (11.4) | 524 (12.9) | 209 (10.2) | 465 (9.0) |
Missing | 1039 (4.6) | 301 (5.0) | 174 (3.6) | 247 (5.7) | 81 (3.8) | 236 (4.4) |
Analgesic use | ||||||
Monthly | 3964 (17.8) | 1044 (17.6) | 728 (15.4) | 726 (17.2) | 322 (15.2) | 1144 (21.5) |
Weekly | 2883 (12.9) | 770 (13.0) | 516 (10.9) | 496 (11.7) | 234 (11.0) | 867 (16.3) |
Daily | 15 471 (69.3) | 4122 (69.4) | 3478 (73.7) | 3007 (71.1) | 1566 (73.8) | 3301 (62.1) |
Missing | 389 (1.7) | 122 (2.0) | 66 (1.4) | 82 (1.9) | 19 (0.9) | 100 (1.9) |
Paresis grade | ||||||
Normal | 18 469 (81.3) | 5217 (86.2) | 3867 (80.8) | 3635 (84.3) | 1480 (69.2) | 4270 (78.9) |
Mild | 2683 (11.8) | 504 (8.3) | 593 (12.4) | 378 (8.8) | 433 (20.3) | 775 (14.3) |
Severe | 1555 (6.9) | 337 (5.6) | 328 (6.9) | 298 (6.9) | 225 (10.5) | 367 (6.8) |
Previous surgery | ||||||
0 | 17 469 (77.5) | 4765 (79.2) | 3626 (75.9) | 3135 (73.5) | 1704 (80.4) | 4239 (79.0) |
1 | 3924 (17.4) | 991 (16.5) | 863 (18.1) | 803 (18.8) | 334 (15.8) | 933 (17.4) |
≥2 | 1160 (5.1) | 261 (4.3) | 291 (6.1) | 330 (7.7) | 82 (3.9) | 196 (3.7) |
Missing | 154 (0.7) | 41 (0.7) | 8 (0.2) | 43 (1.0) | 18 (0.8) | 44 (0.8) |
Microdiscectomy | 21 255 (93.6) | 5523 (91.2) | 4338 (90.6) | 4065 (94.3) | 2114 (98.9) | 5215 (96.4) |
Surgical levels ≥2 | 1263 (5.6) | 466 (7.7) | 246 (5.1) | 124 (2.9) | 91 (4.3) | 336 (6.2) |
Emergency surgery | 4397 (19.5) | 1022 (17.1) | 1366 (28.6) | 1018 (23.7) | 921 (43.3) | 70 (1.3) |
Missing | 131 (0.6) | 68 (1.1) | 12 (0.3) | 16 (0.4) | 10 (0.5) | 25 (0.5) |
Abbreviations: ASA, American Society of Anesthesiologists; BMI, body mass index (calculated as weight in kilograms divided by height in meters squared); NRS, Numeric Rating Scale; ODI, Oswestry Disability Index.
A 10-item score from 0 (none) to 100 (maximum disability) encompassing limitations in various activities of daily living. Treatment success is defined based on achievement of the minimal important change (≥22 points improvement from baseline).
Individuals receiving retirement/age pension. While the retirement age in Norway is 67 years, individuals have the flexibility to decide when they wish to start receiving their retirement pension.
Pending medical or insurance claim or litigation issue.
EQ-5D questionnaire; 5th item, moderate to severe (3L) or moderate to extreme (5L).
The NRS measures pain intensity during the last week on an 11-point scale, with 0 representing no pain and 10 the worst imaginable pain
No features had more than 6% missing values. The proportions with missing outcome data were 35% (ODI) and 37% (NRS back pain and NRS leg pain).
Model Development
The predictive performance of the 7 different machine learning algorithms was compared using estimates of the random effects meta-analysis per algorithm and outcome. The difference between the maximum and minimum C statistic was only 0.01 for each outcome. However, calibration intercepts and slopes varied substantially across the algorithms (eTable 3 in Supplement 1). Extreme gradient boosting had the highest discriminatory performance for each outcome while also showing excellent calibration.
Apparent Predictive Model Performance
The ODI model was able to discriminate between patients with treatment success and nonsuccess with an apparent C statistic of 0.83 (95% CI, 0.82-0.84). The C statistic was 0.78 (95% CI, 0.77-0.78) for NRS back pain and 0.76 (95% CI, 0.76-0.77) for NRS leg pain.
Internal-External Cross-Validation
Model discrimination (C statistic) and calibration metrics (slope and intercept) from internal-external cross-validation of the ODI model are shown in Figure 2. C statistics were similar across regions, with point estimates ranging from 0.81 to 0.84 (pooled random-effects meta-analysis estimate, 0.82; 95% CI, 0.81-0.84). Positive predictive values ranged from 0.81 to 0.88 (pooled estimate, 0.86; 95% CI, 0.82-0.89) and negative predictive values ranged from 0.51 to 0.63 (pooled estimated, 0.58; 95% CI, 0.52-0.64) (eTable 4 in Supplement 1). Calibration slopes were consistent across regions (point estimates, 0.94-1.03; pooled estimate, 0.99; 95% CI, 0.93-1.06). There was minor heterogeneity in calibration intercept across regions, likely due to some variation in outcome incidence between regions (point estimates, −0.05 to 0.11; pooled estimate, 0.01; 95% CI, −0.07 to 0.10). The overall calibration plot for the ODI model is shown in Figure 3A and by region in eFigure 2A in Supplement 1.
Figure 2. Internal-External Cross-Validation in 5 Validation Cohorts and the Overall Estimation Across Validation Cohorts for Oswestry Disability Index.
REML+ HKSJ indicates restricted maximum likelihood + Hartung-Knapp-Sidik-Jonkman.
Figure 3. Assessment of Overall Calibration.
The dashed blue line indicates perfect calibration. The orange line is a fitted Loess smoother curve for the predicted probabilities. NRS indicates Numeric Rating Scale; ODI, Oswestry Disability Index.
For NRS back pain and NRS leg pain, discrimination was somewhat lower, with C statistics ranging from 0.75 to 0.80 (pooled estimate, 0.77; 95% CI, 0.75-0.79) for NRS back pain and 0.74 to 0.77 (pooled estimate, 0.75; 95% CI, 0.74-0.76) for NRS leg pain (eFigure 1 in Supplement 1). Predictive values are reported in eTable 4 in Supplement 1. The calibration slope was also similar across regions, ranging from 0.96 to 1.09 for NRS back pain and 0.91 to 1.10 for NRS leg pain. After meta-analysis, the summary calibration slope was 1.01 (95% CI, 0.94-1.07) for NRS back pain and 1.02 (95% CI, 0.92-1.12) for NRS leg pain. Calibration intercept was consistent across regions for NRS back pain (point estimate, −0.06 to 0.08; pooled estimate, 0.00; 95% CI, −0.07 to 0.08). For NRS leg pain (point estimate, −0.09 to 0.14; pooled estimate, −0.01; 95% CI, −0.14 to 0.11), the overall risk was underestimated in private hospitals (0.14; 95% CI, 0.03-0.25). The NRS back pain and NRS leg pain overall calibration plots are shown in Figure 3B and C and by region in eFigure 2B and C in Supplement 1.
Decision curve analyses in the validation sets from internal-external validation are shown in eFigure 3 in Supplement 1. For all outcomes, the prediction model had higher net benefit than the treat-all or treat-none strategies across a broad range of threshold probabilities.
Feature Importance
The most important features for treatment success on the ODI were higher baseline score of the outcome, shorter duration of back pain, no previous surgery, and no symptoms of anxiety and depression (Figure 4). The same features were also among the most influential for NRS back pain and NRS leg pain (eFigure 4 and eFigure 5 in Supplement 1).
Figure 4. Shapley Additive Explanations (SHAP) Summary Plot of Variable Importance for the Oswestry Disability Index Model.
Predictive features are arranged based on their importance. Each dot represents 1 prediction result. SHAP values indicate the distribution of the prediction among the features; a positive value contributes to treatment success, while a negative value contributes to nonsuccess. ASA indicates American Society of Anesthesiologists; and NRS, Numeric Rating Scale.
Sensitivity Analyses
Consistent results were found for all 3 models in sensitivity analyses including only surgical cases with complete data (n = 11 461 for ODI, n = 11 944 for NRS back pain, and n = 11 321 for NRS leg pain). Calibration slopes, calibration intercept, and C statistics are shown in eFigure 6 in Supplement 1. (eFigure 6 in Supplement 1).
Discussion
In this prognostic study, we developed and validated machine learning models for prediction of treatment success or nonsuccess 12 months after lumbar discectomy. Internal-external cross-validation showed that our models had consistently good calibration when applied to the different geographic regions and private hospitals, and good discrimination with C statistics 0.81 or greater for disability and 0.74 or greater for pain. The models integrated 25 routinely available preoperative features and should be readily implementable in standard clinical settings at the point of surgical decision-making.
The discriminative performance of our models is generally similar to or better than earlier prediction models for disability and pain improvement following lumbar discectomy.10,11,14 Staartjes et al11 developed a deep learning–based ODI prediction model, with a C statistic of 0.84. However, they only included 422 patients from a single-center registry and did not assess calibration. Similarly, Halicka et al14 developed disability and pain prediction models, including both patients with disc herniation and spinal stenosis. Although the models demonstrated good calibration in temporal validation data, the discriminatory ability was acceptable at best (C statistics from 0.62 to 0.72). In contrast, disability and pain prediction models with good discrimination (C statistics from 0.79 to 0.83) have been developed using data from the Danish national registry for spine surgery (DaneSpine).10 DaneSpine and NorSpine are very similar in terms of patient inclusion and data collection processes,7 but the present study is an important extension of this previous work. We used data from the whole NORspine registry (40 centers and 22 707 patients), compared with only patients from a single center of the DaneSpine (n = 1968 patients). We also used internal-external cross-validation to provide insights into heterogeneity and evaluate the generalizability of the models.
In the pursuit to develop the best prediction models, 7 machine learning algorithms were trained and tested. The algorithms showed similar discriminatory ability; however, random forest and extra trees underperformed in terms of calibration with intercepts far below 0 and slopes above 1 across all models (eTable 3 in Supplement 1). Overall, these findings are consistent with a study exploring the added value of machine learning algorithms to regression models for prognostication in traumatic brain injury.39 A 2019 systematic review also found no evidence of superior performance of machine learning over logistic regression in studies with relatively small sample sizes (median n = 1250).40 While machine learning is known to be data hungry and thrive with high-dimensional data,41 we did not find incremental value of our machine learning models compared with logistic regression despite the larger sample size and 25 predictor variables included. However, we emphasize that multiple models should be explored and compared when developing prediction models.
Limitations
Our study has limitations. It was based on a large nationwide spine register using input data that align well with data available in spine registries worldwide, providing unique external validation opportunities; however, there are probably important features that we could not include due to availability, and the incremental predictive value of other predictors should be explored in external validation studies with model updating. Furthermore, the detail of some data types is also suboptimal in the NORspine registry (eg, previous medication, health care use, and work status). Including data from electronic health registries may improve predictive performance. Enrichment of our models with these data types is a subject for further work.
There are other limitations to our study. Although missingness of predictive features was low, the rate of missing outcome data was high. However, we accounted for missing data using a multiple imputation procedure, and complete-case analysis showed consistent results. Analyses from the NORspine registry also indicate that loss to follow-up does not bias conclusions about treatment effects, with no major differences in patient-reported outcomes between nonrespondents and respondents.42,43
Furthermore, a single agreed-on cutoff for defining benefit following lumbar discectomy is yet to be established. We chose estimates of treatment success as the outcome, calculated using anchor-based predictive modeling.24,25 Similar cutoffs for substantial benefit have recently been established in the Canadian Spine registry.44 In addition, our study sample only consisted of patients undergoing surgery (specialist health care), and their potential outcomes following nonsurgical treatment remain unknown. We argue that patients at high risk of not achieving a substantial benefit from surgery should be recommended other treatment pathways, but an impact study is needed to examine potential outcomes following nonsurgical treatment. We also acknowledge that important questions remain regarding the optimal timing of surgery,45 which we were not able to shed light on within our study design.
Conclusion
We developed and validated machine learning models with high to moderate discriminative performance for predicting success or nonsuccess in disability and pain 12 months after lumbar disc herniation surgery. The models were based on routinely available preoperative predictors, making them readily amenable to further external validation in other spine registries and potentially implementable in electronic medical records systems to inform about individual prognosis and aid in surgical decision-making.
eTable 1. Description of Predictors
eTable 2. Hyperparameters of Machine Learning Models
eTable 3. Average Performance Evaluation of Machine Learning Models
eTable 4. Predictive Values Across Validation Cohorts and Overall for Oswestry Disability Index, Numeric Rating Scale Back Pain, and Numeric Rating Scale Leg Pain
eFigure 1. Internal-External Cross-Validation in Five Validation Cohorts and the Overall Estimation Across Validation Cohorts for (A) Numeric Rating Scale Back Pain and (B) Numeric Rating Scale Leg Pain
eFigure 2. Assessment of Calibration in Validation Cohorts for (A) Oswestry Disability Index, (B) Numeric Rating Scale Back Pain, and (C) Numeric Rating Scale Leg Pain
eFigure 3. Decision Curve Analysis During Internal-External Cross-Validation for (A) Oswestry Disability Index, (B) Numeric Rating Scale Back Pain, and (C) Numeric Rating Scale Leg Pain
eFigure 4. SHAP Summary Plot of Variable Importance for Numeric Rating Scale Back Pain
eFigure 5. SHAP Summary Plot of Variable Importance for Numeric Rating Scale Leg Pain
eFigure 6. Internal-External Cross-Validation in Five Validation Cohorts and the Overall Estimation Across Validation Cohorts Including Only Surgical Cases With Complete Data for (A) Oswestry Disability Index, (B) Numeric Rating Scale Back Pain, and (C) Numeric Rating Scale Leg Pain
Data Sharing Statement
References
- 1.Grotle M, Småstuen MC, Fjeld O, et al. Lumbar spine surgery across 15 years: trends, complications and reoperations in a longitudinal observational study from Norway. BMJ Open. 2019;9(8):e028743. doi: 10.1136/bmjopen-2018-028743 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Martin BI, Mirza SK, Spina N, Spiker WR, Lawrence B, Brodke DS. Trends in lumbar fusion procedure rates and associated hospital costs for degenerative spinal diseases in the United States, 2004 to 2015. Spine (Phila Pa 1976). 2019;44(5):369-376. doi: 10.1097/BRS.0000000000002822 [DOI] [PubMed] [Google Scholar]
- 3.Solumsmoen S, Poulsen G, Kjellberg J, Melbye M, Munch TN. The impact of specialised treatment of low back pain on health care costs and productivity in a nationwide cohort. EClinicalMedicine. 2021;43:101247. doi: 10.1016/j.eclinm.2021.101247 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Bailey CS, Rasoulinejad P, Taylor D, et al. Surgery versus conservative care for persistent sciatica lasting 4 to 12 months. N Engl J Med. 2020;382(12):1093-1102. doi: 10.1056/NEJMoa1912658 [DOI] [PubMed] [Google Scholar]
- 5.Weinstein JN, Tosteson TD, Lurie JD, et al. Surgical vs nonoperative treatment for lumbar disk herniation: the Spine Patient Outcomes Research Trial (SPORT): a randomized trial. JAMA. 2006;296(20):2441-2450. doi: 10.1001/jama.296.20.2441 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Gibson JN, Waddell G. Surgical interventions for lumbar disc prolapse: updated Cochrane Review. Spine (Phila Pa 1976). 2007;32(16):1735-1747. doi: 10.1097/BRS.0b013e3180bc2431 [DOI] [PubMed] [Google Scholar]
- 7.Lagerbäck T, Fritzell P, Hägg O, et al. Effectiveness of surgery for sciatica with disc herniation is not substantially affected by differences in surgical incidences among three countries: results from the Danish, Swedish and Norwegian spine registries. Eur Spine J. 2019;28(11):2562-2571. doi: 10.1007/s00586-018-5768-9 [DOI] [PubMed] [Google Scholar]
- 8.Porter ME. What is value in health care? N Engl J Med. 2010;363(26):2477-2481. doi: 10.1056/NEJMp1011024 [DOI] [PubMed] [Google Scholar]
- 9.Rihn JA, Berven S, Allen T, et al. Defining value in spine care. Am J Med Qual. 2009;24(6)(suppl):4S-14S. doi: 10.1177/1062860609349214 [DOI] [PubMed] [Google Scholar]
- 10.Pedersen CF, Andersen M, Carreon LY, Eiskjær S. Applied machine learning for spine surgeons: predicting outcome for patients undergoing treatment for lumbar disc herniation using PRO data. Global Spine J. 2022;12(5):866-876. doi: 10.1177/2192568220967643 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Staartjes VE, de Wispelaere MP, Vandertop WP, Schröder ML. Deep learning–based preoperative predictive analytics for patient-reported outcomes following lumbar discectomy: feasibility of center-specific modeling. Spine J. 2019;19(5):853-861. doi: 10.1016/j.spinee.2018.11.009 [DOI] [PubMed] [Google Scholar]
- 12.McGirt MJ, Sivaganesan A, Asher AL, Devin CJ. Prediction model for outcome after low-back surgery: individualized likelihood of complication, hospital readmission, return to work, and 12-month improvement in functional disability. Neurosurg Focus. 2015;39(6):E13. doi: 10.3171/2015.8.FOCUS15338 [DOI] [PubMed] [Google Scholar]
- 13.Willems SJ, Coppieters MW, Rooker S, Heymans MW, Scholten-Peeters GGM. Baseline patient characteristics commonly captured before surgery do not accurately predict long-term outcomes of lumbar microdiscectomy followed by physiotherapy. Spine (Phila Pa 1976). 2020;45(14):E885-E891. doi: 10.1097/BRS.0000000000003448 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Halicka M, Wilby M, Duarte R, Brown C. Predicting patient-reported outcomes following lumbar spine surgery: development and external validation of multivariable prediction models. BMC Musculoskelet Disord. 2023;24(1):333. doi: 10.1186/s12891-023-06446-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.McGirt MJ, Bydon M, Archer KR, et al. An analysis from the Quality Outcomes Database, part 1: disability, quality of life, and pain outcomes following lumbar spine surgery: predicting likely individual patient outcomes for shared decision-making. J Neurosurg Spine. 2017;27(4):357-369. doi: 10.3171/2016.11.SPINE16526 [DOI] [PubMed] [Google Scholar]
- 16.van Hooff ML, Jacobs WC, Willems PC, et al. Evidence and practice in spine registries. Acta Orthop. 2015;86(5):534-544. doi: 10.3109/17453674.2015.1043174 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Schoenfeld AJ. Spine surgical research: searching for absolute truth in the era of “big data”. Spine J. 2015;15(5):803-805. doi: 10.1016/j.spinee.2015.01.007 [DOI] [PubMed] [Google Scholar]
- 18.Ogink PT, Groot OQ, Bindels BJJ, Tobert DG. The use of machine learning prediction models in spinal surgical outcome: an overview of current development and external validation studies. Semin Spine Surg. 2021;33(2):100872. doi: 10.1016/j.semss.2021.100872 [DOI] [Google Scholar]
- 19.Steyerberg EW, Moons KG, van der Windt DA, et al. ; PROGRESS Group . Prognosis Research Strategy (PROGRESS) 3: prognostic model research. PLoS Med. 2013;10(2):e1001381. doi: 10.1371/journal.pmed.1001381 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Collins GS, Reitsma JB, Altman DG, Moons KG. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. BMJ. 2015;350:g7594. doi: 10.1136/bmj.g7594 [DOI] [PubMed] [Google Scholar]
- 21.Solberg TK, Ingebrigtsen T, Olsen LR, Thyrhaug AM. NORspine annual report. September 8, 2023. Accessed April 10, 2023. https://unn.no/fag-og-forskning/medisinske-kvalitetsregistre/nasjonalt-kvalitetsregister-for-ryggkirurgi#arsrapporter
- 22.Fairbank JC, Pynsent PB. The Oswestry Disability Index. Spine (Phila Pa 1976). 2000;25(22):2940-2952. doi: 10.1097/00007632-200011150-00017 [DOI] [PubMed] [Google Scholar]
- 23.Von Korff M, Jensen MP, Karoly P. Assessing global pain severity by self-report in clinical and health services research. Spine (Phila Pa 1976). 2000;25(24):3140-3151. doi: 10.1097/00007632-200012150-00009 [DOI] [PubMed] [Google Scholar]
- 24.Terluin B, Eekhout I, Terwee CB, de Vet HC. Minimal important change (MIC) based on a predictive modeling approach was more precise than MIC based on ROC analysis. J Clin Epidemiol. 2015;68(12):1388-1396. doi: 10.1016/j.jclinepi.2015.03.015 [DOI] [PubMed] [Google Scholar]
- 25.Terluin B, Eekhout I, Terwee CB. The anchor-based minimal important change, based on receiver operating characteristic analysis or predictive modeling, may need to be adjusted for the proportion of improved patients. J Clin Epidemiol. 2017;83:90-100. doi: 10.1016/j.jclinepi.2016.12.015 [DOI] [PubMed] [Google Scholar]
- 26.Kamper SJ, Ostelo RWJG, Knol DL, Maher CG, de Vet HCW, Hancock MJ. Global perceived effect scales provided reliable assessments of health transition in people with musculoskeletal disorders, but ratings are strongly influenced by current status. J Clin Epidemiol. 2010;63(7):760-766.e1. doi: 10.1016/j.jclinepi.2009.09.009 [DOI] [PubMed] [Google Scholar]
- 27.Riley RD, Ensor J, Snell KIE, et al. Calculating the sample size required for developing a clinical prediction model. BMJ. 2020;368:m441. doi: 10.1136/bmj.m441 [DOI] [PubMed] [Google Scholar]
- 28.Lopez CD, Boddapati V, Lombardi JM, et al. Artificial learning and machine learning applications in spine surgery: a systematic review. Global Spine J. 2022;12(7):1561-1572. doi: 10.1177/21925682211049164 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Riley RD, Van Calster B, Collins GS. A note on estimating the Cox-Snell R2 from a reported C statistic (AUROC) to inform sample size calculations for developing a prediction model with a binary outcome. Stat Med. 2021;40(4):859-864. doi: 10.1002/sim.8806 [DOI] [PubMed] [Google Scholar]
- 30.Mikkelsen E, Ingebrigtsen T, Thyrhaug AM, et al. The Norwegian Registry for Spine Surgery (NORspine): cohort profile. Eur Spine J. 2023;32(11):3713-3730. doi: 10.1007/s00586-023-07929-5 [DOI] [PubMed] [Google Scholar]
- 31.Eddings W, Marchenko Y. Accounting for clustering with mi impute. 2011. Accessed May 1, 2023. https://www.stata.com/support/faqs/statistics/clustering-and-mi-impute/
- 32.Rubin D. Multiple Imputation for Nonresponse in Surveys. Wilson & Sons; 1987. [Google Scholar]
- 33.Steyerberg EW, Harrell FE Jr. Prediction models need appropriate internal, internal-external, and external validation. J Clin Epidemiol. 2016;69:245-247. doi: 10.1016/j.jclinepi.2015.04.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Debray TP, Moons KG, Ahmed I, Koffijberg H, Riley RD. A framework for developing, implementing, and evaluating clinical prediction models in an individual participant data meta-analysis. Stat Med. 2013;32(18):3158-3180. doi: 10.1002/sim.5732 [DOI] [PubMed] [Google Scholar]
- 35.de Jong VMT, Moons KGM, Eijkemans MJC, Riley RD, Debray TPA. Developing more generalizable prediction models from pooled studies and large clustered data sets. Stat Med. 2021;40(15):3533-3559. doi: 10.1002/sim.8981 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.IntHout J, Ioannidis JP, Borm GF. The Hartung-Knapp-Sidik-Jonkman method for random effects meta-analysis is straightforward and considerably outperforms the standard DerSimonian-Laird method. BMC Med Res Methodol. 2014;14(1):25. doi: 10.1186/1471-2288-14-25 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Vickers AJ, Elkin EB. Decision curve analysis: a novel method for evaluating prediction models. Med Decis Making. 2006;26(6):565-574. doi: 10.1177/0272989X06295361 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Lundberg SM, Lee SI. A unified approach to interpreting model predictions. In: Proceedings of the Advances in Neural Information Processing Systems. 2017:4768–4777. [Google Scholar]
- 39.Gravesteijn BY, Nieboer D, Ercole A, et al. ; CENTER-TBI collaborators . Machine learning algorithms performed no better than regression models for prognostication in traumatic brain injury. J Clin Epidemiol. 2020;122:95-107. doi: 10.1016/j.jclinepi.2020.03.005 [DOI] [PubMed] [Google Scholar]
- 40.Christodoulou E, Ma J, Collins GS, Steyerberg EW, Verbakel JY, Van Calster B. A systematic review shows no performance benefit of machine learning over logistic regression for clinical prediction models. J Clin Epidemiol. 2019;110:12-22. doi: 10.1016/j.jclinepi.2019.02.004 [DOI] [PubMed] [Google Scholar]
- 41.Beam AL, Kohane IS. Big data and machine learning in health care. JAMA. 2018;319(13):1317-1318. doi: 10.1001/jama.2017.18391 [DOI] [PubMed] [Google Scholar]
- 42.Solberg TK, Sørlie A, Sjaavik K, Nygaard ØP, Ingebrigtsen T. Would loss to follow-up bias the outcome evaluation of patients operated for degenerative disorders of the lumbar spine? Acta Orthop. 2011;82(1):56-63. doi: 10.3109/17453674.2010.548024 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Kaur S, Alhaug OK, Dolatowski FC, Solberg TK, Lønne G. Characteristics and outcomes of patients who did not respond to a national spine surgery registry. BMC Musculoskelet Disord. 2023;24(1):164. doi: 10.1186/s12891-023-06267-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Power JD, Perruccio AV, Canizares M, et al. Determining minimal clinically important difference estimates following surgery for degenerative conditions of the lumbar spine: analysis of the Canadian Spine Outcomes and Research Network (CSORN) registry. Spine J. 2023;23(9):1323-1333. doi: 10.1016/j.spinee.2023.05.001 [DOI] [PubMed] [Google Scholar]
- 45.Schmid AB, Dove L, Ridgway L, Price C. Early surgery for sciatica. BMJ. 2023;381:791. doi: 10.1136/bmj.p791 [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
eTable 1. Description of Predictors
eTable 2. Hyperparameters of Machine Learning Models
eTable 3. Average Performance Evaluation of Machine Learning Models
eTable 4. Predictive Values Across Validation Cohorts and Overall for Oswestry Disability Index, Numeric Rating Scale Back Pain, and Numeric Rating Scale Leg Pain
eFigure 1. Internal-External Cross-Validation in Five Validation Cohorts and the Overall Estimation Across Validation Cohorts for (A) Numeric Rating Scale Back Pain and (B) Numeric Rating Scale Leg Pain
eFigure 2. Assessment of Calibration in Validation Cohorts for (A) Oswestry Disability Index, (B) Numeric Rating Scale Back Pain, and (C) Numeric Rating Scale Leg Pain
eFigure 3. Decision Curve Analysis During Internal-External Cross-Validation for (A) Oswestry Disability Index, (B) Numeric Rating Scale Back Pain, and (C) Numeric Rating Scale Leg Pain
eFigure 4. SHAP Summary Plot of Variable Importance for Numeric Rating Scale Back Pain
eFigure 5. SHAP Summary Plot of Variable Importance for Numeric Rating Scale Leg Pain
eFigure 6. Internal-External Cross-Validation in Five Validation Cohorts and the Overall Estimation Across Validation Cohorts Including Only Surgical Cases With Complete Data for (A) Oswestry Disability Index, (B) Numeric Rating Scale Back Pain, and (C) Numeric Rating Scale Leg Pain
Data Sharing Statement