Abstract
Purpose
Adolescent idiopathic scoliosis (AIS) affects a significant portion of the adolescent population, leading to severe spinal deformities if untreated. Diagnosis, surgical planning, and assessment of outcomes are determined primarily by the Cobb angle on anteroposterior spinal radiographs. Screening for scoliosis enables early interventions and improved outcomes. However, screenings are often conducted through school entities where a trained radiologist may not be available to accurately interpret the imaging results.
Methods
In this study, we developed an artificial intelligence tool utilizing a keypoint region-based convolutional neural network (KR-CNN) for automated thoracic Cobb angle (TCA) measurement. The KR-CNN was trained on 609 whole-spine radiographs of AIS patients and validated using our institutional AIS registry, which included 83 patients who underwent posterior spinal fusion with both preoperative and postoperative anteroposterior X-ray images.
Results
The KR-CNN model demonstrated superior performance metrics, including a mean absolute error (MAE) of 2.22, mean squared error (MSE) of 9.1, symmetric mean absolute percentage error (SMAPE) of 4.29, and intraclass correlation coefficient (ICC) of 0.98, outperforming existing methods.
Conclusion
This method will enable fast and accurate screening for AIS and assessment of postoperative outcomes and provides a development framework for further automation and validation of spinopelvic measurements.
Keywords: Humans, Spinal curvatures, Adolescent, Artificial intelligence, Spinal fusion, Neural
Introduction
Adolescent idiopathic scoliosis (AIS) affects 2–4% of the adolescent population, causing a three-dimensional deformity of the spinal column that can impair normal motion and posture, potentially leading to lung and heart dysfunction, early-onset osteoarthritis, and disc degeneration if untreated [34, 36]. Spinal fusion surgery in progressive cases remains the main treatment option, but variations in pre-operative characteristics, surgical implants, and maneuvers result in a wide range of outcomes, with approximately 20% being less than satisfactory [7, 35].
Accurate assessment of spinopelvic parameters, particularly the Cobb angle, is essential for optimizing surgical strategies and enhancing patient outcomes [10, 14, 26, 27]. Part of the planning process includes reviewing whole-spine standing radiographs to measure these parameters [9, 13, 23]. However, traditional TCA measurements using coronal radiographs are error-prone due to subjective interpretation, variability among practitioners, and differences in training, necessitating improved methodologies [10, 26, 27]. This variability can lead to imprecise evaluations, highlighting the need for more reliable methods. Moreover, screenings are often conducted at schools, where there is often limited access to radiologists with training in interpreting the relevant images. This could lead to an increased rate of false-positive and false-negative screening tests that could hinder patient outcomes or yield increased healthcare costs [8, 31]. In this study, we investigated the development and validation of a keypoint region-based convolutional neural network (KR-CNN) for the automated measurement of the coronal TCA. KR-CNNs facilitate image analysis by identifying specific points on an image through a multi-step classification task. KR-CNN operate through a multi-step process. Initially, the network applies a series of convolutional layers to extract high-level features from the input image. This is followed by the generation of a region proposal network, in which potential regions of interest are identified where keypoints may be located. These keypoints are specific, identifiable points or features within an image, in this case representing anatomic landmarks. These regions are then refined and classified by the network to precisely define structures. This method enables the accurate identification of recurring patterns and shapes in the setting of variable backgrounds. Consequently, this is particularly well-suited for radiographic measurement tasks in which differences in imaging settings can obscure the identification of well-defined structures. Calculating the TCA angle involves identifying the bony structures of the spine, followed by the extension of lines within a plane. These measurements assist in surgical decision-making and treatment planning for patients with AIS, thereby improving surgical outcomes and patient satisfaction. Overall, the goal of this National Institutes of Health R21-funded feasibility study was to demonstrate the utility of the KR-CNN approach for the fast, accurate, and generalizable measurement of spinopelvic features.
Methods
Guidelines
The methodology and reporting of our study were supported by established guidelines, including the Strengthening the Reporting of Observational Studies in Epidemiology (STROBE), the Transparent Reporting of a Multivariable Prediction Models for Individual Prognosis or Diagnosis + Artificial Intelligence (TRIPOD + AI), and the Checklist for Artificial Intelligence in Medical Imaging (CLAIM) [4, 6, 25].
Data source
The Penn Medicine and Children's Hospital of Philadelphia (CHOP) Institutional Review Board (IRBs 852054, 22–020186) granted approval and authorized the review of Electronic Health Records (EHR). All methods were performed in accordance with relevant guidelines and regulations. Due to the retrospective design of the study, the need for direct patient consent was waived. The database, “Dataset 16,” which was used to train the KR-CNN to automate coronal TCA measurements, was requested from SpineWeb and consisted of 609 anterior–posterior spinal X-ray images annotated with keypoints for Cobb angle determination as part of the Accurate Automated Spinal Curvature Estimation (AASCE) 2019 Grand Challenge [5, 37]. For the validation, an institutional registry for AIS correction surgery was constructed at our center, involving manual chart reviews (MCR) and the integration of medical imaging data from our Picture Archiving and Communication System (PACS). The creation of this registry was a collective endeavor under the multi-institutional umbrella of the Penn Medicine Healthcare System. Prior to the transfer of data from CHOP, all necessary data sharing agreements were established and finalized.
Patient selection and outcomes
Database records of surgeries performed between June 13, 2014, and December 19, 2022, were reviewed, to include pediatric patients (age < 18 years old) who underwent elective thoracolumbosacral posterior spinal fusion (PSF) for AIS correction. Exclusion criteria were age at surgery above 18 years, inclusion of cervical regions in the fusion construct, or surgical intervention performed for any indication other than deformity, such as tumor, trauma, or infection. The primary outcome of this study was the development and validation of a framework to fully automate coronal TCA measurements. Secondary outcomes included comparison of model performance to other models prior reported in the literature.
Missing data
The prevalence of missingness in tabular data was 0%, and a complete-case approach was used. No imputations were performed. All records included in model development and validation had complete imaging data.
Statistics, model development, and external validation
Baseline characteristics were summarized using descriptive statistics. Dataset 16 from SpineWeb was used to train the model. For preprocessing, bounding boxes were derived from keypoint locations, ensuring compatibility with the KR-CNN model. Images were normalized, resized, and transformed into tensors. Low-confidence detections were filtered using confidence thresholding and non-max suppression to retain only relevant instances. The hidden layers used a ResNet + FPN as their backbone. The output layer included information on keypoints, confidence scores, and bounding boxes. Using derived information from the output layer, a function was utilized to automate the calculation of radiographic measurements using extracted coordinates (Fig. 1). Validation was performed using our institutional registry. Both radiographic automated measurements and manual measurements were utilized to calculate performance metrics of the measurement automation tool. These metrics included mean absolute error (MAE), median absolute error (MedAE), mean squared error (MSE), symmetric mean absolute percentage error (SMAPE), and the intraclass correlation coefficient (ICC). Percentile-based bootstrapping with 10,000 resamples was used to calculate 95% confidence intervals for the MedAE. All development, statistical analysis, and the creation of figures were performed using Python, version 3.11 (Python Foundation, Wilmington, DE).
Fig. 1.

Overview of the developed and validated keypoint region-based convolutional neural network (KR-CNN) architecture
Results
Primary outcomes
The KR-CNN with a ResNet-50-FPN backbone was trained using “Dataset 16,” which consists of 609 spinal anterior–posterior (coronal) X-ray images with annotated keypoints needed to calculate the TCA angle. This dataset also includes TCA angles that have been pre-computed on each image for model training. After establishing the model, we assessed its performance on a separate dataset from CHOP. In order to validate the performance of our method on a broad range of local images, our study included 83 patients who underwent multilevel posterior spinal fusion (PSF) for AIS correction. Demographics and baseline characteristics are included in Table 1.
Table 1.
Baseline and surgical characteristics of the validation cohort undergoing long-segment posterior spinal fusion for adolescent idiopathic scoliosis correction
| Variables | Overall (n = 83) |
|---|---|
| Age | 13 (11–14) |
| Female sex | 61 (73.5%) |
| Race | |
| White | 56 (67.5%) |
| Black | 18 (21.7%) |
| Other | 9 (10.8%) |
| Hispanic ethnicity | 6 (7.2%) |
| BMI (kg/m2) | 21.6 (19.4–25.8) |
| Prior non-surgical intervention | 51 (61.4) |
| Radiographic measurements | |
| Right TC | 82 (98.8) |
| TCA (°) | 56.0 (52.0–64.0) |
| Left LC | 82 (98.8) |
| LCA (°) | 42.0 (32.0–52.0) |
| PSF (levels) | 11 (10–12) |
| PSO (levels) | 0 (0–4) |
Data are median (interquartile range) or number of patients (%)
BMI Body-mass-index, TC Thoracic curve, TCA Thoracic cobb angle, LC Lumbar curve, LCA Lumbar cobb angle. PSF Posterior spinal fusion, PSO Pedicle subtraction osteotomy
Radiographic images from our institutional registry used for validation were in the standard Digital Imaging and Communications in Medicine (DICOM) format. Coronal whole-spine standing radiographs had a mean resolution of 4264 × 1980 with a mean dots per inch (dpi) value of 135. Representative pre-operative and post-operative coronal images from one patient, including annotations and predictions are shown (Fig. 2). We assessed the performance of the automated method by comparing the KR-CNN results to values calculated by board-certified radiologists. Our model had a MAE of 2.22 (95% confidence interval: 1.06—3.39), MedAE of 1.47 (0.89–3.15), MSE of 9.1, ICC of 0.98, and SMAPE of 4.29 (Table 2). A regression calibration plot is illustrated (Fig. 3).
Fig. 2.

Whole-spine standing radiographs of a patient undergoing long-segment posterior spinal fusion for adolescent idiopathic scoliosis correction. The images display preoperative coronal annotations (A), postoperative coronal annotations (B), preoperative coronal predictions (C), and postoperative coronal predictions (D), including keypoints and bounding boxes
Table 2.
Performance metrics of the keypoint region-based convolutional neural network (KR-CNN) including external and internal validation for automated measurement of thoracic Cobb angles in the validation cohort of patients undergoing long-segment posterior spinal fusion for adolescent idiopathic scoliosis correction
| Model | MAE ± SD (95% CI) | MedAE (IQR; 95% CI) | MSE | SMAPE | ICC GT |
|---|---|---|---|---|---|
| KR-CNN Resnet 50 (Current Study, External Validation) | 2.22 ± 2.04 (1.06 to 3.39) | 1.47 (0.73–2.97; 0.89 to 3.15) | 9.1 | 4.29 | 0.98 |
| KR-CNN Resnet 50 (Current Study, Internal Validation) | 1.78 ± 1.61 (1.12 to 2.44) | 1.21 (0.65–2.43; 0.81 to 2.74) | 6.84 | 3.67 | 0.99 |
| Landmark Net | 13.85 | — | — | 38.87 | — |
| Angle Net | 9.36 | — | — | 23.91 | — |
| AEC-Net | 4.90 | — | — | 23.59 | — |
| Vertebra Landmark Detection | — | — | — | 10.81 | — |
| Auto-CA | — | — | — | 5.27 | — |
| VLTENet | 2.51 | — | — | 5.44 | — |
| RegNet | 2.92 | — | — | 6.87 | |
| MVE-Net | 7.81 | 24.94 | |||
| Seg4Reg | — | — | — | 21.71 | — |
| ScolioVis | — | — | — | 8.97 | — |
| Minimum bounding rectangle | — | — | — | — | 0.97 |
ICC Intraclass correlation coefficient, GT ground truth, IQR interquartile range, MAE Mean absolute error, MedAE Median absolute error, MSE Mean squared error, KR-CNN Keypoint Region-based convolutional neural network, Resnet 50 residual network 50, SD Standard deviation, SMAPE Symmetric mean absolute percentage error
Fig. 3.
Regression calibration plot of the validation cohort with 95% confidence interval
Secondary outcomes
These results demonstrated a notable improvement in accuracy compared to existing automated TCA measurement methods, including AEC-Net [3], Seg4Reg [19], ScolioVis [30], Auto-CA [28], VLTNet [42] and others [1, 3, 12, 18, 33, 41]. These previous methods utilized a variety of deep learning frameworks including direct Cobb angle regression and methods based on landmark detection and vertebra segmentation. The SMAPE is a commonly reported performance metric for automated measurement methods, and our value of 4.29 is lower than existing methods, outperforming Auto-CA which had the second lowest SMAPE value of 5.27. Similarly, the MAE of 2.22 achieved in this study outperformed the methods that reported this metric, including VLTENet, which achieved the second lowest value of 2.51. Moreover, by demonstrating the performance on external data from the training set, our method uniquely demonstrated that this superior performance metrics are generalizable across institutions, while most methods report validation metrics by splitting a single dataset into training and testing subset. For example, VLTENet, which achieved a SMAPE of 5.44 on a test subset of the AASCE dataset, this value dropped to 13.9 when the method was applied to an external clinical dataset [42]. Taken together, we report a method using a KR-CNN for automated TCA measurement that outperformed existing methods and demonstrated robust accuracy on external data.
Discussion
In this study we developed an KR-CNN for the automated measurement of the thoracic Cobb angle. The benchmarking results of our method indicated that the implementation of this machine learning algorithm yielded precise TCA measurements, demonstrating its feasibility to assist in the screening, evaluation, and early intervention in AIS cases. Previous research has demonstrated the application of convolutional neural networks (CNNs) for automating these measurements with high accuracy and consistency [11, 22]. These studies highlight the potential of AI to overcome the limitations of manual measurement, characterized by high interobserver variability and potential for human error. Recently, there have been advances in R-CNN model implementations that had promising results [20, 21, 30]. However, these methods have not been widely adopted into clinical practice because of their lower accuracy and lack of generalizability. Our study presents a model that provides superior performance and enhanced generalizability. Our framework is adaptable and can be readily trained on a broad range of datasets to optimize the results at the level of specific institutions or even individual users.
The model demonstrated exceptional performance across all metrics, particularly in MAE and SMAPE, which outperformed outcomes reported in recent literature. The high ICC further establishes its reliability. These findings suggest that the model's accuracy and consistency are noteworthy within the current landscape of automated spinal measurements. Before clinical integration, thorough validation is essential, emphasizing the model’s potential to enhance the reliability and efficiency of clinical decision-making processes.
Convolutional neural network approaches hold broad potential for their clinical applicability. Previous studies have demonstrated the utility of R-CNN methods for fast and accurate radiographic measurement and feature detection in a range of contexts, ranging from periodontics [2, 32] to neurosurgery [24, 38] to pathology [29]. In these disparate contexts, AI-assisted image analysis may improve both efficiency and accuracy in tasks involving the identification of structures and feature measurements. Factors such as human error, variability in anatomic landmarks, and differences in imaging quality can yield a striking degree of error in conceptually straightforward tasks. AI technologies can detect small changes in background intensity at the level of individual pixels, rendering these approaches far more sensitive to identify patterns with high confidence that are too subtle or microscopic for a human observer to discern. A strength of CNNs is their ability to define object boundaries and edges at the pixel level. This enables highly consistent measurements of distances and quantification of shape features on radiographic images, abrogating potential concerns for inter-rater reliability. Collectively, improving diagnostic accuracy has far-reaching consequences, including improvements in clinical decision-making, surgical planning, and post-operative follow-up.
Overall, recent advancements in deep learning-based landmark detection have focused on improving accuracy, reliability, and adaptability [16, 17, 40]. New models, such as multi-task learning networks, simultaneously detect key landmarks and segment anatomical structures, reducing the need for multiple separate models [40]. Others, like FDGR-Net, refine landmark detection by filtering out irrelevant background noise and improving focus on subtle anatomical features [17]. Although vision learning models, a type of multimodal deep learning model, may serve a crucial role in medical imaging, there are several deficiencies in clinical applicability, costs, and data sharing and.protection [15]. In spinal imaging, landmark-aware networks enhance Cobb angle measurement by improving feature detection and minimizing the impact of imaging artifacts. These innovations, along with our findings, highlight the growing potential of CNN models to automate medical imaging tasks and integrate seamlessly into clinical workflows.
Limitations
The findings of the present methodological feasibility study should be interpreted in the context of its limitations. Although the model was rigorously trained and validated, its population—AIS patients—may be considered skewed relative to the general population. However, since AIS imaging often presents greater interpretative challenges, particularly due to included hardware, the model is expected to generalize well, performing at least as accurately, if not better, in broader cohorts. Nonetheless, further external validation is necessary to confirm its robustness and clinical applicability. While our imaging dataset is comprehensive, including a broad range of coronal TCA measurements, the inclusion of more images across various populations and retraining the model using United States data, followed by external validation, could improve generalizability. Fine-tuning the KR-CNN with data augmentations could further enhance model robustness. Nonetheless, by using our institutional data, including a wide range of preoperative and postoperative X-ray images, as a testing dataset, we already demonstrated external validity beyond the larger registry of public data that was used to train the model. We acknowledge the critical importance of data governance and recognize that future implementation studies must focus on curating governable datasets that ensure appropriate access and oversight. Additionally, other spinal parameters have been associated with AIS outcomes that could be included in the model, including but not limited to location of apical vertebra/disc, convexity of the curve, thoracic kyphosis, lumbar lordosis, and pelvic tilt [39]. Another limitation is the potential for the model to encounter difficulties with outliers or atypical cases, which may not be adequately represented in the training data. The automated approach, while highly accurate in standard cases, could struggle with images that present unique challenges, such as severe deformities, poor image quality, or unusual anatomical variations. This underscores the need for further studies to assess the model’s performance and suggests that the implementation of a confidence output that could estimate the level of error in the given prediction and flag outliers for manual review would be a beneficial addition in future iterations. Moreover, the utilization of multiple imaging views, in addition to the standard coronal view, could further improve accuracy, especially in outlier cases. This could ultimately be extended to incorporate three-dimensional imaging and longitudinal data. Finally, widespread adoption into clinical practice remains a challenge for any AI model, including the KR-CNN. Factors such as integration with existing clinical workflows, user acceptance, and the need for continuous monitoring and updates of the model as new data becomes available must be carefully considered before widespread clinical application.
Conclusion
Overall, this study highlights the precision and reliability of AI-based automation of spinal imaging interpretation through the development of a KR-CNN for thoracic Cobb angle measurement, while providing a framework for further optimization and application of this dynamic model. Future research should focus on expanding datasets, refining the algorithm, conducting rigorous external validations, clinical impact, and integration. Ultimately, the successful implementation of an automated workflow for spinopelvic measurement has the potential to reduce costs, limit human error, and improve patient outcomes.
Author’s contribution
M.M.D.: conception and design of the study, data acquisition, data analysis, interpretation of data, creation of new software used in the work, manuscript drafting, manuscript revision, and approval of submission. J.H.S.: data acquisition, interpretation of data, manuscript drafting, manuscript revision, and approval of submission. J.G: interpretation of data, manuscript drafting, manuscript revision, and approval of submission. B.R.B.: data acquisition, interpretation of data, manuscript drafting, manuscript revision, and approval of submission. M.K.: interpretation of data, manuscript drafting, manuscript revision, and approval of submission. J.W.Y.: interpretation of data, manuscript drafting, manuscript revision, and approval of submission. A.K.O.: interpretation of data, manuscript drafting, manuscript revision, and approval of submission. P.J.C. data acquisition, interpretation of data, manuscript drafting, manuscript revision, and approval of submission. J.A.: data acquisition, interpretation of data, manuscript drafting, manuscript revision, and approval of submission. B.W..: interpretation of data, manuscript drafting, manuscript revision, and approval of submission. W.C.W.: conception and design of the study, interpretation of data, creation of new software used in the work, manuscript drafting, manuscript revision, and approval of submission.
Funding
This work was partially supported by a grant from the National Institutes of Health, USA (NIH grant R21AR075971) and the Catherine Sharpe Foundation.
Data availability
In accordance with the University of Pennsylvania Institutional Review Board requirements, the research data supporting this project may not be publicly shared. De-identified data may be shared upon reasonable request following execution of a Data Usage Agreement. Please reach out to corresponding author, M.M.D., with any inquiries.
Declarations
Ethical approval
The study was approved by the Penn Medicine and Children's Hospital of Philadelphia (CHOP) Institutional Review Board (IRBs 852054, 22–020186). This study has been performed in accordance with the ethical standards laid down in the 1964 Declaration of Helsinki and its later amendments.
Consent to participate
All patients (or their parent or legal guardian in the case of children under 16) provided written informed consent for inclusion in this study. No organs/tissues were utilized in this study. Only radiographic imaging was utilized in the study.
Consent to publish
The authors affirm that human research participants provided informed consent for publication of the images in Fig. 2.
Competing interests
In the past 36 months, AKO has received consulting fees from Medacta and Johnson and Johnson (with no relation to this work) and has served as an E2M ad-hoc reviewer for the Journal of Neurosurgery Publishing Group (JNS PG). Additionally, within the same period, JWY has received a grant from Johnson and Johnson; consulting fees from Medyssey, TrackX, Richard Wolf, and Johnson and Johnson; holds patents planned, issued, or pending with Kinesiometrics and MedCyclops; and has served in a leadership role on the Scientific Program Committee of the AANS/CNS Joint Section on Disorders of the Spine and Peripheral Nerves (with no relation to this work). All other authors report no conflict of interest.
Footnotes
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Mert Marcel Dagli and Jonathan H. Sussman contributed equally to this work.
References
- 1.Bernstein P, Metzler J, Weinzierl M, Seifert C, Kisel W, Wacker M (2021) Radiographic scoliosis angle estimation: spline-based measurement reveals superior reliability compared to traditional COBB method. Eur Spine J 30:676–685. 10.1007/s00586-020-06577-3 [DOI] [PubMed] [Google Scholar]
- 2.Cha J-Y, Yoon H-I, Yeo I-S, Huh K-H, Han J-S (2021) Peri-implant bone loss measurement using a region-based convolutional neural network on dental periapical radiographs. J Clin Med 10:1009. 10.3390/jcm10051009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Chen B, Xu Q, Wang L, Leung S, Chung J, Li S (2019) An automated and accurate spine curve analysis system. IEEE Access 7:124596–124605. 10.1109/ACCESS.2019.2938402 [Google Scholar]
- 4.Collins GS, Moons KGM, Dhiman P, Riley RD, Beam AL, Calster BV, Ghassemi M, Liu X, Reitsma JB, van Smeden M, Boulesteix A-L, Camaradou JC, Celi LA, Denaxas S, Denniston AK, Glocker B, Golub RM, Harvey H, Heinze G, Hoffman MM, Kengne AP, Lam E, Lee N, Loder EW, Maier-Hein L, Mateen BA, McCradden MD, Oakden-Rayner L, Ordish J, Parnell R, Rose S, Singh K, Wynants L, Logullo P (2024) Tripod+AI statement: updated guidance for reporting clinical prediction models that use regression or machine learning methods. BMJ 385:e078378. 10.1136/bmj-2023-078378 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Constant C, Aubin C-E, Kremers HM, Garcia DVV, Wyles CC, Rouzrokh P, Larson AN (2023) The use of deep learning in medical imaging to improve spine care: a scoping review of current literature and clinical applications. North American Spine Society Journal (NASSJ) 15:100236. 10.1016/j.xnsj.2023.100236 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.von Elm E, Altman DG, Egger M, Pocock SJ, Gøtzsche PC, Vandenbroucke JP, STROBE Initiative (2007) The strengthening the reporting of observational studies in epidemiology (STROBE) statement: guidelines for reporting observational studies. Lancet 370:1453–1457. 10.1016/S0140-6736(07)61602-X [DOI] [PubMed] [Google Scholar]
- 7.Fischer CR, Kim Y (2011) Selective fusion for adolescent idiopathic scoliosis: a review of current operative strategy. Eur Spine J 20:1048–1057. 10.1007/s00586-011-1730-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Fischer CR, Terran J, Lonner B, McHugh B, Warren D, Glassman S, Bridwell K, Schwab F, Lafage V (2014) Factors predicting cost-effectiveness of adult spinal deformity surgery at 2 years. Spine Deform 2:415–422. 10.1016/j.jspd.2014.04.011 [DOI] [PubMed] [Google Scholar]
- 9.Greiner KA (2002) Adolescent idiopathic scoliosis: radiologic decision-making. Am Fam Physician 65:1817–1822 [PubMed] [Google Scholar]
- 10.Gstoettner M, Sekyra K, Walochnik N, Winter P, Wachter R, Bach CM (2007) Inter- and intraobserver reliability assessment of the Cobb angle: manual versus digital measurement tools. Eur Spine J 16:1587–1592. 10.1007/s00586-007-0401-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Harake ES, Linzey JR, Jiang C, Joshi RS, Zaki MM, Jones JC, Khalsa SSS, Lee JH, Wilseck Z, Joseph JR, Hollon TC, Park P (2024) Development and validation of an artificial intelligence model to accurately predict spinopelvic parameters. J Neurosurg Spine. 10.3171/2024.1.SPINE231252 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Horng M-H, Kuok C-P, Fu M-J, Lin C-J, Sun Y-N (2019) Cobb angle measurement of spine from X-ray images using convolutional neural network. Comput Math Methods Med 2019:6357171. 10.1155/2019/6357171 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Kotwicki T (2008) Evaluation of scoliosis today: examination, X-rays and beyond. Disabil Rehabil 30:742–751. 10.1080/09638280801889519 [DOI] [PubMed] [Google Scholar]
- 14.Langensiepen S, Semler O, Sobottke R, Fricke O, Franklin J, Schönau E, Eysel P (2013) Measuring procedures to determine the Cobb angle in idiopathic scoliosis: a systematic review. Eur Spine J 22:2360–2371. 10.1007/s00586-013-2693-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Li X, Li L, Jiang Y, Wang H, Qiao X, Feng T, Luo H, Zhao Y (2025) Vision-language models in medical image analysis: from simple fusion to general large models. Inf Fusion 118:102995. 10.1016/j.inffus.2025.102995 [Google Scholar]
- 16.Li X, Lv S, Li M, Zhang J, Jiang Y, Qin Y, Luo H, Yin S (2023) SDMT: spatial dependence multi-task transformer network for 3D knee MRI segmentation and landmark localization. IEEE Trans Med Imaging 42:2274–2285. 10.1109/TMI.2023.3247543 [DOI] [PubMed] [Google Scholar]
- 17.Li X, Lv S, Zhang J, Li M, Rodriguez-Andina JJ, Qin Y, Yin S, Luo H (2024) Fdgr-net: feature decouple and gated recalibration network for medical image landmark detection. Expert Syst Appl 238:121746. 10.1016/j.eswa.2023.121746 [Google Scholar]
- 18.Liang Y, Lv J, Li D, Yang X, Wang Z, Li Q (2023) Accurate cobb angle estimation on scoliosis X-ray images via deeply-coupled two-stage network with differentiable cropping and random perturbation. IEEE J Biomed Health Inform 27:1488–1499. 10.1109/JBHI.2022.3229847 [DOI] [PubMed] [Google Scholar]
- 19.Lin Y, Zhou HY, Ma K, Yang X, Zheng Y (2020) Seg4Reg networks for automated spinal curvature estimation. In: Cai Y, Wang L, Audette M, Zheng G, Li S (eds) Computational methods and clinical applications for spine imaging. CSI 2019. Springer, Berlin Heidelberg New York, pp 69–74
- 20.Liu Z, Qian L, Jing W, Zhou D, He X, Lou E, Zheng R (2022) Automatic spinal curvature measurement on ultrasound spine images using Faster R-CNN. Arxiv 2204.07988. 10.48550/arXiv.2204.07988
- 21.Ma S, Huang Y, Che X, Gu R (2020) Faster RCNN-based detection of cervical spinal cord injury and disc degeneration. J Appl Clin Med Phys 21:235–243. 10.1002/acm2.13001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Maeda Y, Nagura T, Nakamura M, Watanabe K (2023) Automatic measurement of the Cobb angle for adolescent idiopathic scoliosis using convolutional neural network. Sci Rep 13:14576. 10.1038/s41598-023-41821-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Malfair D, Flemming AK, Dvorak MF, Munk PL, Vertinsky AT, Heran MK, Graeb DA (2010) Radiographic evaluation of scoliosis: review. AJR Am J Roentgenol 194:S8-22. 10.2214/AJR.07.7145 [DOI] [PubMed] [Google Scholar]
- 24.Masood M, Nazir T, Nawaz M, Mehmood A, Rashid J, Kwon H-Y, Mahmood T, Hussain A (2021) A novel deep learning method for recognition and classification of brain tumors from MRI images. Diagnostics 11:744. 10.3390/diagnostics11050744 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Mongan J, Moy L, Kahn CE (2020) Checklist for artificial intelligence in medical imaging (CLAIM): a guide for authors and reviewers. Radiol Artif Intell 2:e200029. 10.1148/ryai.2020200029 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Morrissy RT, Goldsmith GS, Hall EC, Kehl D, Cowie GH (1990) Measurement of the Cobb angle on radiographs of patients who have scoliosis. Evaluation of intrinsic error. J Bone Joint Surg Am 72:320–327 [PubMed] [Google Scholar]
- 27.Pruijs JE, Hageman MA, Keessen W, van der Meer R, van Wieringen JC (1994) Variation in Cobb angle measurements in scoliosis. Skeletal Radiol 23:517–520. 10.1007/BF00223081 [DOI] [PubMed] [Google Scholar]
- 28.Rahmaniar W, Suzuki K, Lin T-L (2024) Auto-CA: automated cobb angle measurement based on vertebrae detection for assessment of spinal curvature deformity. IEEE Trans Biomed Eng 71:640–649. 10.1109/TBME.2023.3313126 [DOI] [PubMed] [Google Scholar]
- 29.Shu T, Shi J, Zheng Y, Jiang Z, Yu L (2023) A key-points based anchor-free cervical cell detector. Annu Int Conf IEEE Eng Med Biol Soc 2023:1–5. 10.1109/EMBC40787.2023.10341092 [DOI] [PubMed] [Google Scholar]
- 30.Taleon C, Elizalde GS, Rubinos CJ (2024) Scoliovis: automated cobb angle measurement of anterior-posterior spine x-rays using multi-instance keypoint detection with keypoint rcnn. https://github.com/Blankeos/scoliovis/blob/main/doc/ScolioVis_Manuscript_IN-EVALUATION.pdf
- 31.Terran J, McHugh BJ, Fischer CR, Lonner B, Warren D, Glassman S, Bridwell K, Schwab F, Lafage V (2014) Surgical treatment for adult spinal deformity: projected cost effectiveness at 5-year follow-up. Ochsner J 14:14–22 [PMC free article] [PubMed] [Google Scholar]
- 32.Vollmer A, Vollmer M, Lang G, Straub A, Kübler A, Gubik S, Brands R, Hartmann S, Saravi B (2023) Automated assessment of radiographic bone loss in the posterior maxilla utilizing a multi-object detection artificial intelligence algorithm. Appl Sci. 10.3390/app13031858 [Google Scholar]
- 33.Wang L, Xu Q, Leung S, Chung J, Chen B, Li S (2019) Accurate automated Cobb angles estimation using multi-view extrapolation net. Med Image Anal 58:101542. 10.1016/j.media.2019.101542 [DOI] [PubMed] [Google Scholar]
- 34.Weinstein SL, Dolan LA, Wright JG, Dobbs MB (2013) Effects of bracing in adolescents with idiopathic scoliosis. N Engl J Med 369:1512–1521. 10.1056/NEJMoa1307337 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Weiss H-R, Bess S, Wong MS, Patel V, Goodall D, Burger E (2008) Adolescent idiopathic scoliosis – to operate or not? A debate article. Patient Saf Surg 2:25. 10.1186/1754-9493-2-25 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Weiss H-R, Karavidas N, Moramarco M, Moramarco K (2016) Long-term effects of untreated adolescent idiopathic scoliosis: a review of the literature. Asian Spine J 10:1163–1169. 10.4184/asj.2016.10.6.1163 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Wu H, Bailey C, Rasoulinejad P, Li S (2017) Automatic Landmark Estimation for Adolescent Idiopathic Scoliosis Assessment Using BoostNet. In: Descoteaux M, Maier-Hein L, Franz A, Jannin P, Collins DL, Duchesne S (eds) Medical Image Computing and Computer Assisted Intervention − MICCAI 2017. Springer International Publishing, Cham, pp 127–135 [Google Scholar]
- 38.Yan J-L, Chen Y-L, Chen M-Y, Chen B-A, Chang J-X, Kao C-C, Hsieh M-C, Peng Y-T, Huang K-C, Chen P-Y (2022) A Robust, Fully Automatic Detection Method and Calculation Technique of Midline Shift in Intracranial Hemorrhage and Its Clinical Application. Diagnostics (Basel) 12:693. 10.3390/diagnostics12030693 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Yang C, Yang M, Chen Y, Wei X, Ni H, Chen Z, Li J, Bai Y, Zhu X, Li M (2015) Radiographic parameters in adult degenerative scoliosis and different parameters between sagittal balanced and imbalanced ADS patients. Medicine (Baltimore) 94:e1198. 10.1097/MD.0000000000001198 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Yang J, Wang J, Meng MQ-H (2024) A landmark-aware network for automated Cobb Angle estimation using x-ray images. Arxiv 2401.12345. 10.48550/arXiv.2405.19645
- 41.Yi J, Wu P, Huang Q, Qu H, Metaxas DN (2020) Vertebra-focused landmark detection for scoliosis assessment. Arxiv 2001.03187. 10.48550/arXiv.2001.03187
- 42.Zou L, Guo L, Zhang R, Ni L, Chen Z, He X, Wang J (2023) vltenet: a deep-learning-based vertebra localization and tilt estimation network for automatic Cobb Angle estimation. IEEE J Biomed Health Inform 27:3002–3013. 10.1109/JBHI.2023.3258361 [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
In accordance with the University of Pennsylvania Institutional Review Board requirements, the research data supporting this project may not be publicly shared. De-identified data may be shared upon reasonable request following execution of a Data Usage Agreement. Please reach out to corresponding author, M.M.D., with any inquiries.

