Skip to main content
PLOS One logoLink to PLOS One
. 2024 May 10;19(5):e0302899. doi: 10.1371/journal.pone.0302899

BACK-to-MOVE: Machine learning and computer vision model automating clinical classification of non-specific low back pain for personalised management

Thomas Hartley 1,#, Yulia Hicks 1,#, Jennifer L Davies 2,3,#, Dario Cazzola 4,5,#, Liba Sheeran 2,3,*,#
Editor: Adrian Pranata6
PMCID: PMC11086851  PMID: 38728282

Abstract

Background

Low back pain (LBP) is a major global disability contributor with profound health and socio-economic implications. The predominant form is non-specific LBP (NSLBP), lacking treatable pathology. Active physical interventions tailored to individual needs and capabilities are crucial for its management. However, the intricate nature of NSLBP and complexity of clinical classification systems necessitating extensive clinical training, hinder customised treatment access. Recent advancements in machine learning and computer vision demonstrate promise in characterising NSLBP altered movement patters through wearable sensors and optical motion capture. This study aimed to develop and evaluate a machine learning model (i.e., ’BACK-to-MOVE’) for NSLBP classification trained with expert clinical classification, spinal motion data from a standard video alongside patient-reported outcome measures (PROMs).

Methods

Synchronised video and three-dimensional (3D) motion data was collected during forward spinal flexion from 83 NSLBP patients. Two physiotherapists independently classified them as motor control impairment (MCI) or movement impairment (MI), with conflicts resolved by a third expert. The Convolutional Neural Networks (CNNs) architecture, HigherHRNet, was chosen for effective pose estimation from video data. The model was validated against 3D motion data (subset of 62) and trained on the freely available MS-COCO dataset for feature extraction. The Back-to-Move classifier underwent fine-tuning through feed-forward neural networks using labelled examples from the training dataset. Evaluation utilised 5-fold cross-validation to assess accuracy, specificity, sensitivity, and F1 measure.

Results

Pose estimation’s Mean Square Error of 0.35 degrees against 3D motion data demonstrated strong criterion validity. Back-to-Move proficiently differentiated MI and MCI classes, yielding 93.98% accuracy, 96.49% sensitivity (MI detection), 88.46% specificity (MCI detection), and an F1 measure of .957. Incorporating PROMs curtailed classifier performance (accuracy: 68.67%, sensitivity: 91.23%, specificity: 18.52%, F1: .800).

Conclusion

This study is the first to demonstrate automated clinical classification of NSLBP using computer vision and machine learning with standard video data, achieving accuracy comparable to expert consensus. Automated classification of NSLBP based on altered movement patters video-recorded during routine clinical examination could expedite personalised NSLBP rehabilitation management, circumventing existing healthcare constraints. This advancement holds significant promise for patients and healthcare services alike.

Introduction

Low back pain (LBP) stands as the most prevalent musculoskeletal ailment globally [1]. It leads to impaired activity and absenteeism, incurring substantial economic consequences [2]. A matter of concern is the escalating growth of disability related to LBP, with the number of years lived with LBP-associated disability surging from 42.5 million in 1990 to 64.9 million in 2017, making LBP the leading global cause of disability [3]. Consequently, LBP has evolved into a prominent global public health concern [4].

Approximately 80–90% of LBP falls under the category of ’non-specific’ (NSLBP), lacking identifiable pathology [5]. Management focuses on reducing pain and disability through active physical approaches involving education, self-management, and exercise tailored to individual needs and capabilities [6]. However, tailoring NSLBP management becomes intricate due to the recognised complexity and heterogenous nature of NSLBP including diverse posture and movement alterations [710], and varying physical conditioning status and disability [11]. This presents a significant clinical hurdle.

Multiple clinical classification systems have emerged to tailor NSLBP management [1214]. Among these, substantial evidence and clinical endorsement exist for multidimensional classification (MDC) framework [13]. The MDC is a multi-stage process, described in detail elsewhere [13, 15], to primarily determine the relative dominance of physical, psychological, behavioural and lifestyle factors and its role in perpetuating the NSLBP disorder to better target management. Individuals whose altered movement and posture patterns contribute to their pain, are classified into either (i) movement impairment (MI), characterised by restricted movement due to pain and avoidance behaviour, or (ii) motor control impairment (MCI), characterised by unrestricted yet painful movement and pain-provoking behaviour [13]. NSLBP patients with MI and MCI, directly amendable through exercise interventions, are believed to represent 60% of all NSLBP cases with MI and MCIs manifesting at different spine levels and movement directions (flexion, extension, side-flexion, rotation), and can also be multi-directional [13].

The distinct nature of the subgroup-specific spine movement and posture alterations was extensively studied and was demonstrated across different NSLBP subgroups during sitting and standing [8, 1619], spine flexion and extension [20, 21] and functional tasks [22]. Further, MDC based targeted NSLBP management, called Cognitive Functional Therapy (CBT), has growing body of high quality evidence of achieving significant and clinically important improvements in patient outcomes compared to standard practice [2326].

Clinical delivery of tailored NSLBP management directly depends on clinicians’ ability classify. Research shows that trained health professionals demonstrate substantial inter-examiner agreement when classifying NSLBP using MDC [15, 27], however, the agreement appears to depend on the level of training received with agreement dropping from Kappa of .90 in those receiving > 100 hours of training to Kappa of .66 in those with < 100 hours of training [15]. Therefore, despite the compelling evidence of its potential clinical benefits, the resource intensive and time-consuming nature of the classification process and the amount of training required makes delivery of tailored NSLBP management immensely challenging [28]. This poses a challenge to widespread clinical adoption.

Machine learning (ML), a subset of artificial intelligence (AI), employs data and algorithms to discern patterns akin to human intelligence. In the medical sphere, ML is increasingly applied for prognosis, diagnosis, and personalised treatments [29]. It has also found utility in diagnosing and prognosticating LBP, though its precision lags behind clinical classification [30]. AI-based classifiers have emerged in studies classifying NSLBP using 3D-motion capture data [18] and wearable sensor data [9, 31]. While their accuracy holds promise, the added equipment costs, maintenance, and significant computational workload impede clinical integration within current healthcare systems.

Computer vision, a facet of AI, extracts profound insights from digital or video imagery. Clinicians routinely use digital images and videos to categorise altered posture and movement patterns associated with NSLBP [8, 15, 27]. Recent strides in computer vision, including convolutional neural networks (CNNs) and expansive training datasets like COCO [32], enabled human pose estimation through video imagery during gait for example [33]. However to our knowledge, no studies have pursued such techniques for NSLBP classification from digital images or video.

In this study, our objective was twofold: Firstly, to construct and evaluate performance of an ML model for NSLBP classification, we called BACK-to-MOVE, to differentiate between the MI and MCI classes using standard video recordings of routine physical examination of NSLBP patients who were clinically classified as either MI or MCI. Given NSLBP’s multifaceted nature, our secondary objective encompassed evaluating the model’s performance in tandem with customary patient-reported outcome measures (PROMs) regularly employed to target NSLBP management.

Methods

The overall protocol utilised in this study consisted of data collection and expert classification of the dataset, pose estimation model design, feature extraction and selection, model training and evaluation described in detail in sections below.

Participants

We used a patient dataset derived from a group of 83 individuals (n = 47 females; mean age [SD] 44.7 [11.8] years; height 170 [9.9] centimetres; mass 81.3 [6.7] kilograms) experiencing NSLBP for over three months. Recruitment was carried out from the physiotherapy waiting lists at Cardiff and Vale Orthopaedic Centre. Eligibility screening was conducted via both telephone and in-person assessments by an experienced physiotherapy clinician (LS), ensuring adherence to inclusion and exclusion criteria (Table 1). Participants attended a single experimental session at the Research Centre for Clinical Kinaesiology, School of Healthcare Sciences, Cardiff University.

Table 1. Eligibility criteria for NSLBP.

Inclusion Criteria
  • Age 18 years +

  • Non-specific low back pain for longer than 3 months

  • Clear mechanical basis for low back pain whereby symptoms are eased and provoked by specific postures or movements based on previously published criteria [13].
    • Movement impairment (MI)–impairment or a deficit is in the movement restriction of the symptomatic spinal level. Pain is triggered by movement of the spine with the movement restricted.
    • Motor control impairment (MCI)—impairment or deficit in the control of the symptomatic spinal level. Pain is triggered by adopting sustained positions such as sitting, bending, standing, twisting, with movement not restricted.
  • Able to speak and understand English well enough to complete questionnaires independently

Exclusion Criteria
  • Primary pain area different to lower back (from Thoracic vertebra T12 to buttock line) e.g., leg or neck pain

  • Acute exacerbation of pain at the time of testing rendering the individual unable to undertake the testing procedure

  • Specific diagnosis for pain (nerve root compression, radicular pain/radiculopathy, disc herniation, spondylolisthesis, spinal stenosis)

  • Surgery (lower limb or abdominal surgery in last 6 months, any spinal surgery)

  • Injection therapy for pain relief in the last 3 months

  • Rheumatologic/inflammatory disease (e.g., psoriatic arthritis, rheumatoid arthritis, ankylosing spondylitis, Scheuermann’s disease) Scoliosis (if a primary pain driver)

  • Progressive neurological or neurodegenerative conditions (e.g. multiple sclerosis, Parkinson’s disease, motor neuron disease)

  • Red flags/serious pathology (malignancy, acute trauma such as fracture, systemic infection, spinal cord compression, cauda equina syndrome.

  • Pregnancy/breast feeding

Data used in this study were collected during a Postdoctoral Research Fellowship (LS) between 1 February 2018 to 31 January 2019 with the archived samples accessed 1 June 2020 to conduct this current study. Health and Care Research Wales Ethics Committee, Wales REC 3 (reference 10/MRE09/28, IRAS project number 51853, amendment number ARUKBBC2017.02) approved the study. Participant anonymity was ensured by applying facial blurring to each frame of the video files, executed by the postdoctoral researcher (TH) using a custom Matlab script. All subsequent data processing and analysis was performed on a fully anonymised dataset.

Clinical classification procedure

Each enrolled NSLBP participant underwent classification utilising the clinical MDC framework [13]. This classification was independently executed by two expert physiotherapists, both well-versed in MDC and trained in accordance with a documented classification protocol [15]. This process entails a subjective evaluation of pain behaviour, scrutiny of available radiological imaging, assessment of pain-related beliefs, lifestyle habits, and physical examination of spinal function. The latter includes gauging the range, speed of movement, and pain response during fundamental movements such as spine flexion, extension and side flexion as well as evaluating slouched and upright sitting postures, and basic functional tasks like sit-to-stand and squat. A clinical judgment is then rendered regarding whether observed movement irregularities are the dominant pain drivers. Subsequently, individuals were broadly categorised into either MI (n = 42) or MCI (n = 41) phenotypes, based on the criteria outlined in Table 1. and independently confirmed by two expert physiotherapists.

Data acquisition

The patient dataset incorporated movement data from a battery of physical assessment of the spinal function informed by expert clinical consensus as integral to NSLBP classification [34]. These assessments encompassed spine flexion, extension, side flexion, along with functional tasks like sit-to-stand, squat and both upright and slouched sitting postures. From this assessment battery, the clinical importance of spine flexion demonstrated the highest expert consensus at 98% [34]. Consequently, spine flexion was chosen as the initial movement test for predicting NSLBP MI and MCI classes in this study. Each participant performed 8–10 repetitions of spine flexion, preceded by two practice repetitions for warm-up. Participants were instructed to begin from a relaxed standing position and bend forward as much as possible, with no further guidance on the method or speed of the movement.

Both an optoelectronic motion capture system (Vicon, Oxford, UK) and a digital video camera (Hero 5, GoPro Inc., CA, USA) captured movement data simultaneously (Fig 1). For motion capture system data, retroreflective markers were positioned on the seventh cervical vertebra (C7), lumbar vertebra (L4), and pelvis (anterior and posterior iliac spines—ASIS, PSIS). Inertial measurement units (not discussed here) attached to the spine necessitated positioning the L4 marker parallel to the fourth lumbar vertebra. Retroreflective marker positions were recorded at 120 Hz using Vicon MX cameras and Vicon Nexus software.

Fig 1. Experimental set up for obtaining spine flexion.

Fig 1

All videos were pre-processed by undergoing rescaling to ensure the shortest side was 768 pixels in length and were resampled for 30 frames per second playback using open-source Python and OpenCV software packages.

Patient-reported outcome measures

Prior to engaging in any physical movement tests, participants were requested to complete PROMs validated and routinely used in LBP to profile pain intensity, levels of disability, fear of movement, pain related catastrophising, self-efficacy, coping strategies, and level of risk of developing persistent LBP disability (Table 2).

Table 2. Patient-reported outcome measures (PROMs).

Measure Domain Description
Visual Analog Scale (VAS) [35] Pain Intensity Level of pain marked on a 100-mm scale anchored with "No pain" and "Worst imaginable pain" at the respective ends.
Oswestry Disability Index (ODI) [36] Disability Participants asked to tick statements that best applying to them in 8 activity of daily living domains (e.g., lifting, personal care and walking). This index’s score spans from zero to 100, with higher scores signifying more pronounced disability.
Tampa Scale of Kinesiophobia [37] Fear of movement Measure of movement-related apprehension in LBP individuals. The scale’s score range is 17 to 68, with elevated scores reflecting greater kinesiophobia.
Pain Catastrophising Scale (PCS) [38] Pain catastrophising 13 statements describing thoughts associated with pain, scored using a 5-point scale anchored with ‘not at all’ and ‘All the time’ at the respective ends.
Pain Self-efficacy Questionnaire (PSEQ) [39] Pain Self-efficacy 10 statements describing activities of daily life scored on a 6-point scale anchored with ‘Not at all confident’ to ‘Completely confident’ at the respective ends.
Coping Strategies Questionnaire (CSQ) [40] Coping A questionnaire asking participants to score 14 items describing different ways of coping with pain using a 6-point scale anchored with ‘Never do that’ to ‘Always do that’ at the respective ends.
STarTBacK tool (SBT) [41] Risk of developing persistent LBP-related disability 9-item screening questionnaire to stratify LBP patients based on their risk (low, medium, high) for persistent LBP-related disability.

Human pose estimation

Convolutional Neural Networks (CNNs) [42]—a subset of artificial neural networks—were harnessed to deduce human body posture from visual data, enabling the phenotyping of NSLBP. Multiple CNN architectures tailored for human pose estimation were considered: CrowdPose [43], DensePose [44], PoseNet [45], UniPose [46] and HigherHRNet [47]. After initial exploration, the HigherHRNet model was chosen due to its exceptional performance, effectively accommodating the varied forward bending movement characteristics inherent in the patient dataset. A detailed comparison of HigherHRNet with alternative CNN architectures was presented in the original article clearly demonstrating its superiority in accurate estimation of the human pose in the videos [47], which was the main influence of the choice.

HigherHRNet, operating as a bottom-up pose estimation model, directly infers human body posture by identifying key body points (keypoints) throughout the entire image [47]. In the initial phase, all 17 keypoints within the HigherHRNet model were identified across the video data of each participant. However, for classifying this NSLBP patient dataset, only specific keypoints—the ankle, hip, and neck, as illustrated in Fig 2, were deemed essential to capture clinically pertinent features essential for classification. The x, y coordinates of each identified keypoint were recorded in pixel space.

Fig 2. Human pose estimation model utilised to estimate spine flexion performance.

Fig 2

By identifying keypoints across all frames within each video, a movement pattern of the participants was established. This concept is exemplified in Fig 3, depicting the x, y coordinates, representing anteroposterior and vertical directions, respectively, of the keypoint representing the neck for a single participant executing multiple repetitions of spine flexion.

Fig 3. Sagittal 2D position of the base of the neck keypoint.

Fig 3

Y = base of the neck vertical coordinate (pixels), X = base of the neck antero-posterior coordinate (pixels).

The initial pre-trained human pose estimation model, HigherHRNet, was trained on the MS-COCO dataset containing 250,000 person instances distributed across 200,000 images [32]. Fine-tuning the pre-trained model wasn’t pursued, as it already exhibited strong performance. This decision aimed to prevent excessive overfitting to the training data, which might compromise the model’s capacity to encapsulate the diverse movement nuances present in the dataset [48].

Feature extraction

In this study, the selected features were deliberately aligned with what holds significance for classifying NSLBP in a clinical context. This approach aimed to avoid overfitting, a well-recognised challenge associated with constructing classifiers that incorporate numerous high-dimensional features [49]. Among the most prevalent NSLBP characteristics are impairments in spatial and temporal parameters of range of movement [7] a focal point of attention during NSLBP assessments undertaken by physiotherapists.

Range of movement was quantified in the form of a spinal angle, computed as the Euclidean distances (in pixels) between three points derived from the pose estimation process. These distances encompass those between the hip and neck keypoints (hn), ankle and hip keypoints (ah), and ankle and neck keypoints (an), as visualised in Fig 4. Armed with these distances, the angle of the spine flexion (θ) was determined using the following formula:

θ=arccoshn2+ah2an2/2hnah

Fig 4. Visual representation of the calculation to obtain the range of movement of spine flexion.

Fig 4

Euclidean distance (in pixels) between hip to neck (hn), ankle to hip (ah), and ankle to neck (an).

Subsequently, a moving average with a 30-frame window (equivalent to 1 second of video data) was employed to smoothen the spine flexion angle waveform. This process aimed to mitigate any unforeseen data fluctuations (Fig 5).

Fig 5. Spine flexion angle waveform of a single participant repeating the movement with the signal pre-smoothing (blue) and post-smoothing (orange).

Fig 5

The spine flexion repetition time (tri), and angle depth (depthi) are highlighted in the graph to give insight into the consistency of participants attainment of a specific flexion depth and the extent to which the flexion depth fluctuated across repetitions.

A collection of both established and new statistical features, commonly applied in clinical movement and NSLBP classification analyses [8, 20, 50], were then computed for the angle signal. These features were employed to assess their predictive value for distinguishing MI and MCI classes within NSLBP. Among the standard features routinely used in clinical movement analysis were the mean, range, minimum, maximum values, variance, and standard deviation for the spine flexion angle θ (Fig 4).

The new features introduced here aimed to evaluate the temporal and spatial consistency of the spinal flexion movement. To achieve this, the angle waveform was scrutinised (Fig 5), with particular emphasis on its minima points, which hold clinical significance as they correspond to the points when participants reach the lowest flexion (Figs 4 and 5). The spine flexion repetition time (tri) was deduced from the frame count between these minima points on the angle graph. Detecting these minima points was accomplished through signal inversion coupled with a peak detection algorithm, in this instance, the approach by Du et al. (2006) [51], implemented through the SciPy python library’s find_peaks_cwt function.

Novel features derived from these minima points encompassed the mean and variance of repetition time (tri) and the variance of the spine flexion angle depth (depthi) in Fig 5, further elaborated below. The variance in depth (depthi) provided insight into the consistency of participants attainment of a specific flexion depth and the extent to which the flexion depth fluctuated across repetitions. Lastly, movement stability, the final feature extracted from the angle graphs, was obtained by dividing the angle waveform (Fig 5) into two halves and determining the angle range for each half, then subtracting one from the other. This feature offered insight into whether a patient increased their maximum bend angle across several repetitions [51].

Considering the multifaceted character of NSLBP, the collected PROMs were incorporated as other important clinical features. This culminated in a set of 16 features for distinguishing between NSLBP MI and MCI subsets (Table 3).

Table 3. Movement and clinical features used for distinguishing between MI and MCI.

Feature Domain Input features
Spine Flexion Angle Standard 1. Variance
2. Standard Deviation
3. Minimal (full spine flexion angle)
4. Range
5. Maximal (spine standing angle)
Novel 6. Repetition Time Mean
7. Repetition Time Variance
8. Depth Variance
9. Movement Stability
Patient Reported Outcome Measures (PROMs) 10. Visual Analog Scale (VAS)
11. Oswestry Disability index (ODI)
12. Tampa Scale of Kinesiophobia (TSK)
13. Pain Catastrophising Scale (PCS)
14. Patient Self-efficacy Questionnaire (PSEQ)
15. Coping Strategies Questionnaire (CSQ)
16. STarTBacK tool (SBT)

(MI = movement impairment, MCI = motor control impairment)

Predictive model design

Category prediction was executed using feedforward neural network, with parameter values drawn from a training dataset comprising labelled instances [52]. The predictive model adopted here (Fig 6) encompassed multiple layers, each featuring a distinct type of input–angle and PROMs (Table 3). These inputs were processed through individual streams, which first underwent batch normalisation [53] before progressing to a feedforward layer. This feedforward layer entailed no hidden nodes; input and output nodes equated the number of features. A Rectified Linear Unit (ReLU) activation function followed the linear layer [54]. The streams were then concatenated and fed into a concluding feedforward layer equipped with a threshold, yielding a binary output (0 or 1) signifying the prediction of either MI or MCI. Throughout this process, model was trained using binary cross-entropy loss and stochastic gradient descent, following a supervised methodology [55].

Fig 6. Predictive model.

Fig 6

Batch Norm = batch normalisation layer, ReLU = Rectified Linear Unit layer.

Evaluation

Pose estimation validation

Criterion validity of the pose estimation model was assessed by comparing the spine flexion angle acquired through pose estimation with the synchronised ’ground truth’ 3-dimensional (3-D) motion capture (Mocap) data, utilising 75% of the NSLBP dataset (n = 62). In the case of the 3-D Mocap spine flexion angle, the pose estimation keypoints (neck and hip) were replaced with the corresponding Mocap positions of C7 and a midpoint between the pelvic markers (PSIS and ASIS). Approximating the ankle’s position involved setting the y position of the L4 marker to zero, generating a virtual ankle marker (vA) consistently positioned directly beneath the L4 marker at a vertical position of zero. The spine flexion angle calculation was then executed as detailed in section on ‘Feature extraction’ above. Subsequently, the pose’s mean square error score was computed to quantify the disparity between the spine flexion angle extrapolated from the Mocap data and that deduced from the pose estimation keypoints.

Predictive model evaluation

The evaluation of the predictive model involved two aspects: i) selecting the most pertinent features and ii) assessing the final model’s performance. To appraise the predictive model’s performance, classification accuracy, along with sensitivity, specificity, and the F1 measure, were chosen as evaluation metrics. Accuracy signifies the fraction of predictions the model correctly attributed. Sensitivity and specificity offer insights into the nature of misclassifications.

In the context of a binary predictive model, as in this study, sensitivity denotes the model’s ability to accurately identify all instances of the positive class (MI), while specificity gauges the model’s capacity to correctly recognise all instances of the negative class (MCI). These metrics were computed employing true positive detections (TP), true negative detections (TN), false positive detections (FP), and false negative detections (FN), using the subsequent formulae:

Accuracy=TP+TNTP+FP+TN+FN
Sensitivity=TPTP+FN
Specificity=TNTN+FP

To gauge the overall classification performance of the predictive model, the F1 measure was computed. The F1 measure represents a balanced combination of precision (sensitivity) and recall (ability to detect class samples), ranging from 0 to 1, where 0 signifies the lowest and 1 indicates the highest overall classification performance [56].

The model assessment occurred in three progressive stages, with outcomes from each stage guiding the subsequent one. The initial stage involved exploring the significance of various feature types in accurate predictions. To achieve this, the model underwent training using all available features, followed by training with a single type of feature. The resulting metric values served as baseline for comparison with subsequent experiment results. In the second stage, the model’s performance was assessed for all permutations of each feature type. Angle features presented 512 permutations [29], while PROM features featured 128 permutations [27]. The final stage encompassed constructing an optimal feature set, in the sense of producing the highest F1-measure among all possible permutations of features in the corresponding model assessment, finalising model training, and executing an ablation experiment to analyse each feature’s impact on the final model’s accuracy. All assessment measures were calculated based on 5-fold cross-validation.

Statistical significance of the results can be evaluated when comparing various competing classification models; however, there is no prior work against which our proposed model could be benchmarked. Nonetheless, the statistical significance of the accuracy obtained in this study was assessed following the method outlined in Combrisson et al., 2015 [57] suitable for comparisons of comparable sample size. For c classes and an infinite number of equally distributed samples, the percent theoretical chance level of classification is 100/c. In our case, the number of classes c = 2, and the chance level of classification is 50%. However, for a limited number of samples, the chance level will depend on it. As Combrisson et al, we assume that the classification errors obey a binomial cumulative distribution, where the probability to predict the correct class at least z times by chance is given by:

Pz=i=znni×1ci×c1cn1

Following this, we used the MATLAB (Mathworks Inc., MA, USA) code provided by Combrisson et al (2015) to compute the statistically significant threshold St(α) = binoinv(1-α, n, 1/c) × 100/n, where n is the sample size, c is the number of classes, α = z/n is the significance level and binoinv() is a MATLAB function [57].

Results

Pose estimation model validity

An illustration of an angle waveform obtained from the pose estimation model and a waveform derived from motion capture data for a single participant performing spine flexion is presented in Fig 7. The mean square error score, calculated to quantify the divergence between the two data sets, was 0.35 degrees. This result showcases robust criterion validity and an adequate level of accuracy between the pose estimation model and the motion capture data.

Fig 7. An example of an angle waveform from a single participant performing 10 repetitions of spine flexion derived from the pose estimation model and the motion capture data (Mocap).

Fig 7

Blue = pose estimation model; Orange = Mocap.

Predictive model evaluation

The classification outcomes from the initial stage of experiments involving all angle and all PROMs input features are presented in Table 4. The higher level of accuracy, sensitivity, and specificity was attained by employing ’all angle features’. The PROMs demonstrated relatively weak capability to precisely identify MCI, resulting in a reduced overall accuracy and implying limited contribution to accurately differentiate between the two NSLBP subsets (Table 4).

Table 4. Classification performance when all spine flexion angle and patient reported outcome measures (PROMs) features were inputted.

Features Accuracy Sensitivity* Specificity** F1 Score
Spine Flexion Angle (all) 87.95% 92.98% 76.92% .914
PROMs (all) 68.67% 91.23% 18.52% .800

*Sensitivity denotes accuracy of correctly assigning Movement Impairment.

**Specificity denotes accuracy of correctly assigning Motor Control Impairment.

F1 = overall classification performance, 0 = lowest, 1 = highest.

Table 5 presents a set of eight optimal input features (marked by X) that was curated from the relevant 16 features on basis of yielding the highest F1 measure value in correctly discriminating between MI and MCI. Interestingly, standard features employed in clinical movement analysis (angle minimum and variance) did not exhibit utility in distinguishing between the subgroups and STarTBacK Tool, a measure of risk in developing disabling LBP, was the only PROM feature selected (Table 5).

Table 5. Set of optimal input features yielding the highest classification performance.

Feature domain Input features Optimal features marked by X
Spine Flexion Angle Standard 1. Variance -
2. Standard Deviation X
3. Minimal (full spine flexion angle) -
4. Range X
5. Maximal (spine standing angle) X
Novel 6. Repetition Time Mean X
7. Repetition Time Variance X
8. Depth Variance X
9. Movement Stability X
Patient Reported Outcome Measures (PROMs) 10. Visual Analog Scale (VAS) -
11. Oswestry Disability index (ODI) -
12. Tampa Scale of Kinesiophobia (TSK) -
13. Pain Catastrophising Scale (PCS) -
14. Patient Self-efficacy Questionnaire (PSEQ) -
15. Coping Strategies Questionnaire (CSQ) -
16. STarTBacK tool (SBT) X

X = features achieving highest F1 measure value among all possible permutations of features

Lastly, the classification performance using solely the optimal input features, both individually and in combination, is depicted in Table 6. The highest overall classification performance of 93.98% was achieved when exclusively utilising optimal angle features. The addition of optimal PROMs feature to the optimal angle features did not yield improvement in the classification performance.

Table 6. Classification performance when the optimal features were inputted separately and combined.

Features Accuracy Sensitivity* Specificity** F1 Score
Optimal Spine Flexion Angle 93.98% 96.49% 88.46% .957
Optimal PROMs 73.49% 91.23% 34.62% .825
Optimal Spine Flexion Angle + PROMs 92.77% 94.74% 88.46% .947

*Sensitivity denotes accuracy of correctly assigning Movement Impairment.

**Specificity denotes accuracy of correctly assigning Motor Control Impairment.

F1 = overall classification performance, 0 = lowest, 1 = highest

The above results were subjected to statistical testing described earlier in the article, achieving statistical significance (p < 0.001) for accuracies of 87.5% and above. Therefore, our results are confirmed to be statistically significant.

Discussion

This is a novel study that examines the feasibility of BACK-to-MOVE; a machine learning and computer vision model for automating clinical classification of NSLBP on basis of altered spine movement captured on standard video camera. A recent systematic review revealed that computer vision’s primary use in LBP predominantly revolves around extracting features from radiological imaging including MRI, CT, X-ray, or ultrasound to identify and/or categorise vertebral structures and intervertebral discs’ pathology levels [30]. While these models may demonstrate clinical applications for specific LBP with a clear patho-anatomical cause, it’s important to note that over 80% of LBP cases are non-specific with clinical guidance actively discouraging the use of imaging to guide its management. [58]. This study stands as the first to employ AI in classifying non-specific LBP utilising video imagery with the intent to aid tailored management of NSLBP as endorsed by clinical guidelines [6]. This approach therefore holds a significant potential to enhance the clinical viability and usefulness of such methods.

This study BACK-to-MOVE classifier achieved percentage accuracy of 93.98% in predicting MI and MCI with F1 score of 0.957, signifying exceptional performance. This level of accuracy favourably compares to inter-examiner agreement of highly experienced physiotherapists undergoing 100+ hours of training to achieve a Kappa coefficient of .82 and percentage agreement of 86%, and surpasses inter-examiner agreement of less experienced clinicians (undergoing less than 100 hours training) achieving Kappa of 0.66 [15]. Our classifier also out-performed clinicians classifying NSLBP from videos of physical examination to reach a Kappa between .55 and .71, and percentage agreement between 65 and 78% [27].

The BACK-to-MOVE classifier showed better prediction of MI class (96.49% sensitivity) than MCI class (88.46% specificity). This difference could stem from the fact that MI, characterised by spine movement restriction, is perhaps more easily discernible from video compared to the potentially subtler deviations in motor control associated with MCI class [10, 13]. Interestingly, Fersum et al. (2009) demonstrated higher clinical agreement classifying MCI (99% agreement) than MI (75% agreement) [15]. This incongruity might be attributed to the uneven distribution of cases in the Fersum et al (2009) study with 24 cases of MCI and one case of MI, along with the clinicians’ suggested unfamiliarity with the less frequently encountered MI [13, 15]. In our study, an almost equal numbers of cases were included across MI (n = 41) and MCI (n = 42).

The highest accuracy (93.98%) was achieved by using seven optimal movement features that characterise the spatial and temporal parameters of spine flexion, extracted from the video imagery. This is in line with recent systematic review and meta-analysis demonstrating that people with LBP have impairments in movement range, speed and spine flexion variability [59]. The advantage of these movement features likely arises from their objective nature encompassing statistical properties of the spine flexion angle’s behaviour across the movement repetitions. Among the 9 movement features, 2 (variance and minimum angle value) were not chosen by the model. This could be because they were already represented by the remaining features, specifically, standard deviation, range, and maximum.

To address the multifactorial nature of NSLBP, we examined the inclusion of seven patient-reported outcome measures (PROMs) typically considered during LBP classification including levels of pain, disability, fear of movement, catastrophising, coping, self-efficacy and risk of chronicity. Using all these PROMs as input features alone achieved low accuracy of 68.67%, rising marginally to 73.39% when only the optimal PROMs were used (STarTBacK). Combining optimal PROMs with the optimal movement features enhanced the classification accuracy to 92.77%, however, this was still lower than inputting the optimal movement features alone (93.98%). The limited discriminative capability of PROMs in our study was likely due to the specific recruitment criteria, which focused on participants with LBP persisting for more than three months, while excluding those experiencing acute and severe LBP. This targeted selection criteria likely contributed to the uniformity of PROMs scores, with variations minimised through excluding the acute and severe cases. In summary, while PROMs undoubtedly provide more holistic picture of the patient’s pain experience, they appear non-discriminatory when it comes to classifying NSLBP of similar intensity and duration based on physical characteristics thus may not be useful features for ML classifiers helping to discern between altered movement patterns.

The BACK-to-MOVE classifier developed in this study utilised a dataset of 83 NSLBP patients, a sample size comparable and exceeding many other studies [60]. To maintain the dataset’s full complexity and minimise selection bias, we included video data from all participants, regardless of challenges like lighting variations. This approach reduces the risk of overestimating predictions by avoiding oversimplification of classification scenarios, an aspect highlighted in previous research [61]. However, some selection bias was unavoidable as we excluded two participant sets with videos too brief for clinicians to classify from. It’s also important to note that participants in our dataset were wearing sports clothing, which could have positively affected pose estimation accuracy. Further exploration is warranted to assess the extent of this influence. Lastly, while participants in our dataset had Vicon markers attached to their spine and pelvis, the model used in this study was trained on the standard COCO dataset, which does not contain any markers. This disparity is therefore unlikely to have significantly impacted pose estimation accuracy.

Strengths and limitations

This study’s primary strength lies in highly innovative use of the state-of-the-art human pose estimation model, HigherHRNet, to classify NSLBP according to a clinical classification model, achieving an accuracy exceeding 93%, which is comparable to clinical expert agreement. Although various pose estimation algorithms and models exist, none have proven reliable for estimating the entire spine’s pose. Therefore, we focused on measuring a single spine flexion angle, which proved effective. However, this single-angle focus might be a limitation considering the complexity of NSLBP demonstrating movement and motor control alterations [7, 8, 10].

Another strength is our use of data from a standard video camera alone. Most clinical research relies on specialised motion capture hardware, not easily accessible in standard clinical settings. Our approach offers an alternative path for potential clinical adoption without high costs or space requirements. Nonetheless, this method’s accuracy depends on video quality and can be influenced by lighting and background conditions.

We utilised movement features already considered important by clinicians trained in NSLBP classification [34]. This will enable easy interpretation by clinicians and immediate opportunities for existing interventions to be tailored to patient needs and capabilities in accordance current clinical guidelines [62]. We also confirmed that these clinician-considered movement features achieve highly accurate automatic NSLBP classification.

A limitation is evaluating a single task (spine flexion) when considering NSLBP heterogeneity in movement and motor control alteration across different tasks [8, 10, 22]. However, recent data indicate that NSLBP individuals exhibit consistent spinal movement patterns across various tasks, known as a spinal movement signature [7]. While this supports focused data collection, discriminating between directional impairment patterns would likely require data from other movements. This study’s feasibility in detecting MI and MCI subsets from a single movement test suggests future research can enhance the automated classifier’s sensitivity across other clinical assessment tests, enabling more tailored rehabilitation approaches when and where this is level of specificity is deemed clinically important (e.g. guide specific exercise protocols for NSLBP individuals in higher risk occupations or sports involving spine loading in specific directions).

Clinical implications

Before considering potential clinical implications, it is important to recognise that classification of a complex and multifactorial such as NSLBP goes beyond physical examination. Other key information including patient history, injury details, radiology (if available), pain behaviour, coping and lifestyle factors are important to identify the dominant pain drivers and target management [24]. Although this study BACK-to-MOVE classifies NSLBP on basis of patient physical function, the primary focus of NSLBP management revolves around restoring physical function [5]. Hence, automating classification of physical aspects is of clinical importance not only to target management but assist its delivery and offer means of objective progression monitoring, thus creating a potential for substantial clinical value to many.

Secondly, NSLBP movement and motor control impairments are described to exhibit directional biases (e.g., extension, flexion, side-flexion, or multi-directional) [13]. In this study, we focused on the initial classification step, determining whether NSLBP primarily involves impaired movement (MI) or impaired motor control (MCI). This distinction enables tailoring the exercise regimen to restore either movement or motor control impairment, as needed.

Self-management tailored to individual needs and capabilities, as endorsed by clinical guidelines, is a cornerstone of managing long-term musculoskeletal pain conditions like NSLBP [6]. Presently, self-management interventions heavily rely on substantial clinical involvement, in-person assessments, and follow-up consultations, with limited avenues for ongoing physical assessment, evaluation, or progress tracking. The BACK-to-MOVE classifier developed in this study offers a significant breakthrough and potential to revolutionise the delivery of tailored self-management interventions. This encompasses automating clinical movement assessment from videos and presents opportunities like remote condition monitoring and providing prompt progress feedback to patients.

Moreover, BACK-to-MOVE holds the potential to address the challenge of offering timely access to personalised, top-quality rehabilitation for all, aligning with the World Health Organization’s call for such advancements [63].

Future research

Future research directions should focus on overcoming the limitations linked to the simplified model of the human spine employed in this study. With advancements in computer vision models, upcoming investigations will delve into movement patterns spanning the entire spinal curvature, capitalising on the clinically significant movement features examined in this study. This holds the potential to yield fresh insights into the development, progression, and subsequent treatment options for movement disorders in the spine. Another avenue for future research involves leveraging our existing data to explore the directional nature of MI and MCI impairments and their correlation with NSLBP-related disability, thus refining rehabilitation strategies. Lastly, expanding the dataset to include both spinal kinematics and muscle electromyography (EMG) will be helping the model detect non-physiological muscle activation further contributing to crafting more precise and clinically meaningful classification models.

Conclusion

This study presents the first automated clinical classification model designed for non-specific low back pain, capable of discerning between clinically recognised movement and motor control impairments in the spine from 2D videos, using computer vision and machine learning. Utilising video data alone, BACK-to-MOVE achieved classification accuracy exceeding 93%, aligning to expert clinician consensus, showcasing its potential for practical clinical application. The automation of NSLBP classification through video-recorded altered movement patterns from routine clinical examinations has the potential to enhance the efficiency of personalised NSLBP rehabilitation management, addressing current healthcare limitations. This advancement holds substantial promise for both patients and healthcare services.

Data Availability

Due to ethical considerations regarding privacy, the video data utilized in this study cannot be shared publicly as they may contain potentially identifying and sensitive information, such as tattoos, birthmarks, sex, ethnicity, and unique movement characteristics. Complete de-identification, noise addition, or blocking of portions of the dataset while preserving its usefulness to replicate study results is not feasible. However, external requests for access to the data can be directed to Professor Valerie Sparkes at arthritiscentre@cardiff.ac.uk.

Funding Statement

This research was funded by the Wellcome Trust [grant number AC1550IF02]. Funding was awarded to the whole team: LS (PI), TH,YH,DC,JD (Co-Is). Funder website: https://wellcome.org/grant-funding/funded-people-and-projects/institutional-strategic-support-fund. The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. For the purpose of Open Access, the author has applied a CC BY public copyright licence to any Author Accepted Manuscript version arising from this submission. This research was part funded by Biomechanics and Bioengineering Research Centre Versus Arthritis (grant number CA20781).

References

  • 1.Hoy D, Bain C, Williams G, March L, Brooks P, Blyth F, et al. A systematic review of the global prevalence of low back pain. Arthritis Rheum. 2012;64(6):2028–37. Epub 2012/01/09. doi: 10.1002/art.34347 . [DOI] [PubMed] [Google Scholar]
  • 2.Krismer M, van Tulder M. Low back pain (non-specific). Best Practice & Research Clinical Rheumatology. 2007;21(1):77–91. doi: 10.1016/j.berh.2006.08.004 [DOI] [PubMed] [Google Scholar]
  • 3.James SL, Abate D, Abate KH, Abay SM, Abbafati C, Abbasi N, et al. Global, regional, and national incidence, prevalence, and years lived with disability for 354 diseases and injuries for 195 countries and territories, 1990–2017: a systematic analysis for the Global Burden of Disease Study 2017. The Lancet. 2018;392(10159):1789–858. doi: 10.1016/S0140-6736(18)32279-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Buchbinder R, van Tulder M, Öberg B, Costa LM, Woolf A, Schoene M, et al. Low back pain: a call for action. The Lancet. 2018;391(10137):2384–8. doi: 10.1016/S0140-6736(18)30488-4 [DOI] [PubMed] [Google Scholar]
  • 5.Maher C, Underwood M, Buchbinder R. Non-specific low back pain. The Lancet. 2017;389(10070):736–47. doi: 10.1016/S0140-6736(16)30970-9 [DOI] [PubMed] [Google Scholar]
  • 6.NICE. National Institute of Health and Care Excellence, Low back pain and Sciatica over 16s: Assessment and Management (NICE guideline NG59)2016 (last updated 2020). https://www.nice.org.uk/guidance/ng59.
  • 7.Christe G, Aussems C, Jolles BM, Favre J. Patients With Chronic Low Back Pain Have an Individual Movement Signature: A Comparison of Angular Amplitude, Angular Velocity and Muscle Activity Across Multiple Functional Tasks. Frontiers in Bioengineering and Biotechnology. 2021;9. doi: 10.3389/fbioe.2021.767974 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Sheeran L, Sparkes V, Caterson B, Busse-Morris M, van Deursen R. Spinal Position Sense and Trunk Muscle Activity During Sitting and Standing in Nonspecific Chronic Low Back Pain Classification Analysis. Spine. 2012;37(8):E486–E95. doi: 10.1097/BRS.0b013e31823b00ce [DOI] [PubMed] [Google Scholar]
  • 9.Laird RA, Keating JL, Kent P. Subgroups of lumbo-pelvic flexion kinematics are present in people with and without persistent low back pain. BMC Musculoskelet Disord. 2018;19(1):309. Epub 2018/08/28. doi: 10.1186/s12891-018-2233-1 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.van Dieën JH, Reeves NP, Kawchuk G, van Dillen LR, Hodges PW. Analysis of Motor Control in Patients With Low Back Pain: A Key to Personalized Care? J Orthop Sports Phys Ther. 2019;49(6):380–8. Epub 2018/06/12. doi: 10.2519/jospt.2019.7916 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Heneweer H, Staes F, Aufdemkampe G, van Rijn M, Vanhees L. Physical activity and low back pain: a systematic review of recent literature. Eur Spine J. 2011;20(6):826–45. Epub 2011/01/09. doi: 10.1007/s00586-010-1680-7 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Foster NE, Hill JC, O’Sullivan P, Hancock M. Stratified models of care. Best Pract Res Clin Rheumatol. 2013;27(5):649–61. doi: 10.1016/j.berh.2013.10.005 . [DOI] [PubMed] [Google Scholar]
  • 13.O’Sullivan P. Diagnosis and classification of chronic low back pain disorders: Maladaptive movement and motor control impairments as underlying mechanism. Manual Therapy. 2005;10(4):242–55. doi: 10.1016/j.math.2005.07.001 [DOI] [PubMed] [Google Scholar]
  • 14.Fersum KV, Dankaerts W, O’Sullivan PB, Maes J, Skouen JS, Bjordal JM, et al. Integration of sub-classification strategies in RCTs evaluating manual therapy treatment and exercise therapy for non-specific chronic low back pain (NSCLBP): a systematic review. Br J Sports Med. 2009. doi: 10.1136/bjsm.2009.063289 . [DOI] [PubMed] [Google Scholar]
  • 15.V Fersum K, O’Sullivan PB, Kvale A, Skouen JS. Inter-examiner reliability of a classification system for patients with non-specific low back pain. Manual Therapy. 2009;14(5):555–61. doi: 10.1016/j.math.2008.08.003 [DOI] [PubMed] [Google Scholar]
  • 16.Astfalck RG, O’Sullivan PB, Smith AJ, Straker LM, Burnett AF. Lumbar spine repositioning sense in adolescents with and without non-specific chronic low back pain—an analysis based on sub-classification and spinal regions. Man Ther. 2013;18(5):410–7. Epub 2013/03/23. doi: 10.1016/j.math.2013.02.005 . [DOI] [PubMed] [Google Scholar]
  • 17.Astfalck RG, O’Sullivan PB, Straker LM, Smith AJ, Burnett A, Caneiro JP, et al. Sitting postures and trunk muscle activity in adolescents with and without nonspecific chronic low back pain: An analysis based on subclassification. Spine. 2010;35(14):1387–95. doi: 10.1097/BRS.0b013e3181bd3ea6 [DOI] [PubMed] [Google Scholar]
  • 18.Sheeran L, Sparkes V, Whatling G, Biggs P, Holt C. Identifying non-specific low back pain clinical subgroups from sitting and standing repositioning posture tasks using a novel Cardiff Dempster–Shafer Theory Classifier. Clinical Biomechanics. 2019;70:237–44. doi: 10.1016/j.clinbiomech.2019.10.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Elgueta-Cancino E, Sheeran L, Salomoni S, Hall L, Hodges PW. Characterisation of motor cortex organisation in patients with different presentations of persistent low back pain. European Journal of Neuroscience. 2021;54(11):7989–8005. doi: 10.1111/ejn.15511 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Dankaerts W, O’Sullivan P, Burnett A, Straker L, Davey P, Gupta R. Discriminating healthy controls and two clinical subgroups of nonspecific chronic low back pain patients using trunk muscle activation and lumbosacral kinematics of postures and movements: A statistical classification model. Spine. 2009;34(15):1610–8. doi: 10.1097/BRS.0b013e3181aa6175 [DOI] [PubMed] [Google Scholar]
  • 21.Dankaerts W, O’Sullivan P. The validity of O’Sullivan’s classification system (CS) for a sub-group of NS-CLBP with motor control impairment (MCI): Overview of a series of studies and review of the literature. Manual Therapy. 2011;16(1):9–14. doi: 10.1016/j.math.2010.10.006 [DOI] [PubMed] [Google Scholar]
  • 22.Hemming R, Sheeran L, van Deursen R, Sparkes V. Non-specific chronic low back pain: differences in spinal kinematics in subgroups during functional tasks. European Spine Journal. 2017. doi: 10.1007/s00586-017-5217-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.O’Sullivan K, Dankaerts W, O’Sullivan L, O’Sullivan PB. Cognitive Functional Therapy for Disabling, Nonspecific Chronic Low Back Pain: Multiple Case-Cohort Study. Physical Therapy. 2015. doi: 10.2522/ptj.20140406 [DOI] [PubMed] [Google Scholar]
  • 24.O’Sullivan PB, Caneiro JP, O’Keeffe M, Smith A, Dankaerts W, Fersum K, et al. Cognitive Functional Therapy: An Integrated Behavioral Approach for the Targeted Management of Disabling Low Back Pain. Physical Therapy. 2018;98(5):408–23. doi: 10.1093/ptj/pzy022 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Vibe Fersum K, Smith A, Kvåle A, Skouen JS, O’Sullivan P. Cognitive functional therapy in patients with non-specific chronic low back pain—a randomized controlled trial 3-year follow-up. European Journal of Pain. 2019;23(8):1416–24. doi: 10.1002/ejp.1399 [DOI] [PubMed] [Google Scholar]
  • 26.O’Keeffe M, O’Sullivan P, Purtill H, Bargary N, O’Sullivan K. Cognitive functional therapy compared with a group-based exercise and education intervention for chronic low back pain: a multicentre randomised controlled trial (RCT). British journal of sports medicine. 2020;54(13):782–9. doi: 10.1136/bjsports-2019-100780 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Dankaerts W, O’Sullivan PB, Straker LM, Burnett AF, Skouen JS. The inter-examiner reliability of a classification method for non-specific chronic low back pain patients with motor control impairment. Man Ther. 2006;11(1):28–39. doi: 10.1016/j.math.2005.02.001 [DOI] [PubMed] [Google Scholar]
  • 28.Sheeran L, Coales P, Sparkes V. Clinical challenges of classification based targeted therapies for non-specific low back pain: What do physiotherapy practitioners and managers think? Manual Therapy. 2014. doi: 10.1016/j.math.2014.11.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Parashar G, Chaudhary A, Rana A. Systematic Mapping Study of AI/Machine Learning in Healthcare and Future Directions. SN computer science. 2021;2(6):461-. Epub 2021/09/16. doi: 10.1007/s42979-021-00848-6 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.D’Antoni F, Russo F, Ambrosio L, Vollero L, Vadalà G, Merone M, et al. Artificial Intelligence and Computer Vision in Low Back Pain: A Systematic Review. International Journal of Environmental Research and Public Health [Internet]. 2021; 18(20). doi: 10.3390/ijerph182010909 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Bacon Z, Hicks Y, Al-Amri M, Sheeran L. Automatic Low Back Pain Classification Using Inertial Measurement Units: A Preliminary Analysis. Procedia Computer Science. 2020;176:2822–31. doi: 10.1016/j.procs.2020.09.272 [DOI] [Google Scholar]
  • 32.Lin T-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, et al., editors. Microsoft COCO: Common Objects in Context. Computer Vision–ECCV 2014; 2014 2014//; Cham: Springer International Publishing. [Google Scholar]
  • 33.Anthony C, Ankit P, Tasos K, Cotton RJ. Portable in-clinic video-based gait analysis: validation study on prosthetic users. medRxiv. 2022:2022.11.10.22282089. doi: 10.1101/2022.11.10.22282089 [DOI] [Google Scholar]
  • 34.Sheeran L, Robling M. Spinal function assessment and exercise performance framework for low back pain. Orthopaedic Proceedings. 2019;101-B(SUPP_9):40-. doi: 10.1302/1358-992x.2019.9.040 [DOI] [Google Scholar]
  • 35.Ferreira-Valente MA, Pais-Ribeiro JL, Jensen MP. Validity of four pain intensity rating scales. PAIN®. 2011;152(10):2399–404. doi: 10.1016/j.pain.2011.07.005 [DOI] [PubMed] [Google Scholar]
  • 36.Fairbank JC, Pynsent PB. The Oswestry Disability Index. Spine. 2000;25:2940–52. doi: 10.1097/00007632-200011150-00017 [DOI] [PubMed] [Google Scholar]
  • 37.Roelofs J, Goubert L, Peters ML, Vlaeyen JWS, Crombez G. The Tampa Scale for Kinesiophobia: further examination of psychometric properties in patients with chronic low back pain and fibromyalgia. European Journal of Pain. 2004;8(5):495–502. doi: 10.1016/j.ejpain.2003.11.016 [DOI] [PubMed] [Google Scholar]
  • 38.Sullivan MJL, Bishop SR, Pivik J. The Pain Catastrophizing Scale: Development and validation. Psychological Assessment. 1995;7:524–32. doi: 10.1037/1040-3590.7.4.524 [DOI] [Google Scholar]
  • 39.Nicholas MK. The pain self-efficacy questionnaire: Taking pain into account. European Journal of Pain. 2007;11(2):153–63. doi: 10.1016/j.ejpain.2005.12.008 [DOI] [PubMed] [Google Scholar]
  • 40.Harland NJ, Georgieff K. Development of the Coping Strategies Questionnaire 24, a Clinically Utilitarian Version of the Coping Strategies Questionnaire. Rehabilitation Psychology. 2003;48(4):296–300. [Google Scholar]
  • 41.Hill JC, Dunn KM, Lewis M, Mullis R, Main CJ, Foster NE, et al. A Primary Care Back Pain Screening Tool: Identifying Patient Subgroups for Initial Treatment. Arthritis Care & Research. 2008;59:632–41. doi: 10.1002/art.23563 [DOI] [PubMed] [Google Scholar]
  • 42.Gong W, Zhang X, Gonzàlez J, Sobral A, Bouwmans T, Tu C, et al. Human Pose Estimation from Monocular Images: A Comprehensive Survey. Sensors (Basel, Switzerland). 2016;16(12):1966. doi: 10.3390/s16121966 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Li J, Wang C, Zhu H, Mao Y, Fang H, Lu C. CrowdPose: Efficient Crowded Scenes Pose Estimation and a New Benchmark. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2019:10855–64.
  • 44.Güler RA, Neverova N, Kokkinos I, editors. DensePose: Dense Human Pose Estimation in the Wild. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2018 18–23 June 2018.
  • 45.Kendall A, Grimes M, Cipolla R, editors. PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization. 2015 IEEE International Conference on Computer Vision (ICCV); 2015 7–13 Dec. 2015.
  • 46.Artacho B, Savakis AE. UniPose: Unified Human Pose Estimation in Single Images and Videos. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2020:7033–42.
  • 47.Cheng B, Xiao B, Wang J, Shi H, Huang TS, Zhang L. HigherHRNet: Scale-Aware Representation Learning for Bottom-Up Human Pose Estimation. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2020:5385–94.
  • 48.Kumar A, Raghunathan A, Jones R, Ma T, Liang P. Fine-Tuning can Distort Pretrained Features and Underperform Out-of-Distribution. ArXiv. 2022;abs/2202.10054. [Google Scholar]
  • 49.França RP, Borges Monteiro AC, Arthur R, Iano Y. Chapter 3—An overview of deep learning in big data, image, and signal processing in the modern digital age. In: Piuri V, Raj S, Genovese A, Srivastava R, editors. Trends in Deep Learning Methodologies: Academic Press; 2021. p. 63–87. [Google Scholar]
  • 50.Hemming R, Sheeran L, van Deursen R, Sparkes V. Non-specific chronic low back pain: differences in spinal kinematics in subgroups during functional tasks. European Spine Journal. 2018;27(1):163–70. doi: 10.1007/s00586-017-5217-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Du P, Kibbe WA, Lin SM. Improved peak detection in mass spectrum by incorporating continuous wavelet transform-based pattern matching. Bioinformatics. 2006;22(17):2059–65. doi: 10.1093/bioinformatics/btl355 [DOI] [PubMed] [Google Scholar]
  • 52.Yadav SS, Jadhav SM. Deep convolutional neural network based medical image classification for disease diagnosis. Journal of Big data. 2019;6(1):1–18. [Google Scholar]
  • 53.Ioffe S, Szegedy C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. 2015/06/01: PMLR. p. 448–56. [Google Scholar]
  • 54.Glorot X, Bordes A, Bengio Y. Deep Sparse Rectifier Neural Networks. 2011/06/14: PMLR. p. 315–23. [Google Scholar]
  • 55.Ian Goodfellow YBaAC. Deep Learning: MIT Press; 2016. [Google Scholar]
  • 56.Taha AA, Hanbury A. Metrics for evaluating 3D medical image segmentation: analysis, selection, and tool. BMC medical imaging. 2015;15:29-. doi: 10.1186/s12880-015-0068-x . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Combrisson E, Jerbi K. Exceeding chance level by chance: The caveat of theoretical chance levels in brain signal classification and statistical assessment of decoding accuracy. Journal of Neuroscience Methods. 2015;250:126–36. doi: 10.1016/j.jneumeth.2015.01.010 [DOI] [PubMed] [Google Scholar]
  • 58.Hall AM, Aubrey-Bassler K, Thorne B, Maher CG. Do not routinely offer imaging for uncomplicated low back pain. BMJ. 2021;372:n291. doi: 10.1136/bmj.n291 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Laird RA, Gilbert J, Kent P, Keating JL. Comparing lumbo-pelvic kinematics in people with and without back pain: a systematic review and meta-analysis. BMC Musculoskeletal Disorders. 2014;15(1):229. doi: 10.1186/1471-2474-15-229 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Amorim P, Paulo J, Silva PA, Peixoto P, Castelo-Branco M, Martins H. Machine Learning Applied to Low Back Pain Rehabilitation–A Systematic Review. International Journal of Digital Health. 2021;1:10. doi: 10.29337/ijdh.34 [DOI] [Google Scholar]
  • 61.Janitza S, Hornung R. On the overestimation of random forest’s out-of-bag error. PLOS ONE. 2018;13(8):e0201904. doi: 10.1371/journal.pone.0201904 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Excellence NIoHaC. Low back pain and Sciatica over 16s: Assessment and Management (NICE guideline NG59)2016 (last updated 2020). https://www.nice.org.uk/guidance/ng59.
  • 63.Mishra S, Gosling J, Laplante-Lévesque A, Zapata T, Azzopardi Muscat N. The need for rehabilitation services in the WHO European Region is substantial and growing. The Lancet Regional Health—Europe. 2023;24:100550. doi: 10.1016/j.lanepe.2022.100550 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Adrian Pranata

20 Feb 2024

PONE-D-23-28397BACK-to-MOVE: Machine Learning and Computer Vision Model Automating Clinical Classification of Non-Specific Low Back Pain for Personalised Management.PLOS ONE

Dear Dr. Sheeran,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Apr 05 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Adrian Pranata, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Note from Emily Chenette, Editor in Chief of PLOS ONE, and Iain Hrynaszkiewicz, Director of Open Research Solutions at PLOS: Did you know that depositing data in a repository is associated with up to a 25% citation advantage (https://doi.org/10.1371/journal.pone.0230416)? If you’ve not already done so, consider depositing your raw data in a repository to ensure your work is read, appreciated and cited by the largest possible audience. You’ll also earn an Accessible Data icon on your published paper if you deposit your data in any participating repository (https://plos.org/open-science/open-data/#accessible-data).

3. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

4. We note that you have indicated that there are restrictions to data sharing for this study. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For more information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. 

Before we proceed with your manuscript, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., a Research Ethics Committee or Institutional Review Board, etc.). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of recommended repositories, please see

https://journals.plos.org/plosone/s/recommended-repositories. You also have the option of uploading the data as Supporting Information files, but we would recommend depositing data directly to a data repository if possible.

We will update your Data Availability statement on your behalf to reflect the information you provide.

5. When completing the data availability statement of the submission form, you indicated that you will make your data available on acceptance. We strongly recommend all authors decide on a data sharing plan before acceptance, as the process can be lengthy and hold up publication timelines. Please note that, though access restrictions are acceptable now, your entire data will need to be made freely accessible if your manuscript is accepted for publication. This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If you are unable to adhere to our open data policy, please kindly revise your statement to explain your reasoning and we will seek the editor's input on an exemption. Please be assured that, once you have provided your new statement, the assessment of your exemption will not hold up the peer review process.

Additional Editor Comments:

Thank you for your patience with the review process. This study is a good step forward in the assessment of movement in people with low back pain. Please address the reviewers' comments on the technical aspect of the study's machine learning algorithm. Thank you for your submission.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The paper has few issues that need further clarifications as follows:

- It is not clear how to validate or ensure data MI and MCI at the beginning before this data fed to feature extraction process.

- The structure of the CNN is not very clear such is the dimension of the input and output.

- Table 5 shows set of optimal input feature yielding the highest classification performance. Instead of using 'X' the actual metric such as classification should be used. Moreover, how to ensure the value is optimal. Also, few input features such as variance, minimal (full spine flexion angle), VAS, ODI, TSK PCS, PSEQ and CSQ do not show as optimal feature. This need to be clarified.

- Since the paper proposed using CNN, there is no information about the training of the CNN; how to avoid overfitting issue, how to ensure each of the training parameter of the CNN is optimal.

- Please explore whether the result is statistically significant.

Reviewer #2: This study aimed to develop and evaluate a machine learning model called "BACK-to-MOVE" for classifying non-specific low back pain (NSLBP) using standard video data, spinal motion data, and patient-reported outcome measures. The model utilized a Convolutional Neural Network architecture for effective pose estimation from video data. Results showed that the "BACK-to-MOVE" model successfully differentiated motor control impairment (MCI) and movement impairment (MI) classes with 93.98% accuracy, demonstrating strong criterion validity. The study concludes that automated classification of NSLBP using computer vision and machine learning could expedite personalized NSLBP rehabilitation management, potentially benefiting patients and healthcare services.

Experiment:

lack comparison among multiple CNNs performance on human pose detection (CrowPose, DensePose, PoseNet, UniPose, HigherHRNet), better provide a simple comparison.

Discussion:

Table 6, the model performance decreases when adding PROMs, this requires more in-depth discussion and ablation study to illustrate that whether this is caused by model. That is whether design a good model the results will change.

Minor error:

Line 76: MDC classification -> MDC, since C means classification.

The label of x and y-axis in Fig. 3 are too vague.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 May 10;19(5):e0302899. doi: 10.1371/journal.pone.0302899.r002

Author response to Decision Letter 0


4 Apr 2024

To PLOS1 Editor:

1. We note that you have indicated that there are restrictions to data sharing for this study. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., a Research Ethics Committee or Institutional Review Board, etc.). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

Our response:

Data sharing: Ethical restrictions exist related to the nature of the data (videos) containing potentially identifying and sensitive information (e.g. tattoos, birth marks, sex, ethnicity, unique movement characteristics etc) with no possibility to de-identify, add noise, or block portions of the dataset and still preserve the usefulness of the dataset to replicate the study results. Unfortunately, our institution (Cardiff University) does not have an established point of contact to field external requests for access to sensitive data.

Reviewer #1:

1. It is not clear how to validate or ensure data MI and MCI at the beginning before this data fed to feature extraction process.

MI/MCI classification is described in methods (clinical classification process lines 147-158). In short, MI/MCI was performed using clinical judgement-based MDC method, which is the only available method and considered as the gold standard. In our study, two MDC-trained expert clinicians independently classified the participants into MI and MCI following a process shown to have above 90% inter-examiner agreement [reference 15 and 27 in reference list]. We added additional detail to clarify the independent confirmation by the 2 experts (line 157).

2. The structure of the CNN is not very clear such is the dimension of the input and output.

The structure of the HigherHRNet CNN, along with a detailed comparative analysis of HigherHRNet performance with that of alternative CNNs for human pose estimation and an ablation study, has been described previously and is referenced in our study [reference 47]. As the architecture of HigherHRNet is not the subject of this study, we limit ourselves to citing this work in our article. Nonetheless, to ensure clarity we added a sentence (line 199-201) referring readers to this detail.

3. Table 5 shows set of optimal input feature yielding the highest classification performance. Instead of using 'X' the actual metric such as classification should be used Moreover, how to ensure the value is optimal. Also, few input features such as variance, minimal (full spine flexion angle), VAS, ODI, TSK PCS, PSEQ and CSQ do not show as optimal feature. This need to be clarified.

The features were optimal in the sense of producing the highest F1 measure value among all possible permutations of features, so using a numerical value against a single feature would not be possible as it is a metric for a feature subset. We added a sentence in the article (line 345-6) and a legend to Table 5 (line 387) to clarify the meaning.

4. Since the paper proposed using CNN, there is no information about the training of the CNN; how to avoid overfitting issue, how to ensure each of the training parameter of the CNN is optimal.

As already detailed in the manuscript (line 220-24) we utilised a CNN model (the selection of which is covered in our response to your comment 2 above) that was pre-trained on the MS-COCO dataset with fine-tuning not pursued given its superior performance and to prevent excessive overfitting.

5. Please explore whether the result is statistically significant.

Thank you for your comment. The aim of this study was to test the performance of the proposed classifier against a clinical judgement classification using MDC method as the ‘gold standard’. The conventional method for this form of performance evaluation is a calculation of accuracy, sensitivity and specificity, and F1-measure (Taha et al 2015, referenced in manuscript). There is no other model against which our proposed model could be benchmarked using statistical testing. However, to respond to your question, we explored the statistical significance of our results following a method outlined by Combrisson et al 2015 [reference 57 added], which showed that our results would reach statistical significance (p-value <0.001) for accuracy of 87.5% and higher. We added this detail and a reference in the article (added lines in methods 350-364, and results 405-407).

Reviewer #2:

1. Experiment: lack comparison among multiple CNNs performance on human pose detection (CrowPose, DensePose, PoseNet, UniPose, HigherHRNet), better provide a simple comparison.

Thank you for your comment. A detailed comparative analysis of the performance of HigherHRNet with that of alternative CNNs for human pose estimation and an ablation study is described in detail in reference [47]. As CNNs performance comparison was not the aim of this study, we limit ourselves to referencing it [47] in our article. Nonetheless, to ensure clarity we added a sentence (line 199-201) referring readers to this detail.

2. Discussion: Table 6, the model performance decreases when adding PROMs, this requires more in-depth discussion and ablation study to illustrate that whether this is caused by model. That is whether design a good model the results will change.

Thank you. In response to your comment, further text has been added to elaborate on the possible reasons for the limited discriminate ability of PROMs (line 463-70). In summary, PROMs in this study measure levels of pain intensity, disability, and psychological distress that vary according to different levels of pain intensity and duration. As participants with LBP > 3 months with pain intensity not precluding them from performing the movements were included, PROMs scores were unlikely to vary across the subgroups. Nevertheless, given the multi-factorial nature of LBP, we felt it was still important to explore any such variation as highlighted in the discussion section.

3. Minor error:

Line 76: MDC classification -> MDC, since C means classification.

Thank you, ‘classification’ has now been removed.

The label of x and y-axis in Fig. 3 are too vague.

Figure 3 has been re-generated with Y axis titled as ‘Base of the neck vertical coordinate (pixels) and X axis titled ‘Base of the neck antero-posterior coordinate (pixels) for improved clarity.

Attachment

Submitted filename: Response to Reviewers.docx

pone.0302899.s001.docx (28.2KB, docx)

Decision Letter 1

Adrian Pranata

16 Apr 2024

BACK-to-MOVE: Machine Learning and Computer Vision Model Automating Clinical Classification of Non-Specific Low Back Pain for Personalised Management.

PONE-D-23-28397R1

Dear Dr. Sheeran,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Adrian Pranata, Ph.D

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Dear Liba,

Thank you for addressing the comments from the reviewers and your patience. Research using AI in low back pain clinical management is scant. This study will add to the emerging area of AI for movement assessment and ultimately, management of people with low back pain. Thank you for your contribution in this important area of research.

Kind regards,

Adrian

Reviewers' comments:

Acceptance letter

Adrian Pranata

29 Apr 2024

PONE-D-23-28397R1

PLOS ONE

Dear Dr. Sheeran,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Adrian Pranata

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response to Reviewers.docx

    pone.0302899.s001.docx (28.2KB, docx)

    Data Availability Statement

    Due to ethical considerations regarding privacy, the video data utilized in this study cannot be shared publicly as they may contain potentially identifying and sensitive information, such as tattoos, birthmarks, sex, ethnicity, and unique movement characteristics. Complete de-identification, noise addition, or blocking of portions of the dataset while preserving its usefulness to replicate study results is not feasible. However, external requests for access to the data can be directed to Professor Valerie Sparkes at arthritiscentre@cardiff.ac.uk.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES