Skip to main content
PLOS One logoLink to PLOS One
. 2021 Jun 3;16(6):e0251892. doi: 10.1371/journal.pone.0251892

Reliability, validity and discriminability of patient reported outcomes for non-specific low back pain in a nationwide physical therapy registry: A retrospective observational cohort study

Guus A Meerhoff 1,2,*, Arie C Verburg 1, Renske M Schapendonk 1,3, Juliette Cruijsberg 1, Maria W G Nijhuis-van der Sanden 1, Simone A van Dulmen 1, Philip J Van der Wees 1
Editor: Alison Rushton4
PMCID: PMC8174721  PMID: 34081704

Abstract

Background

A national clinical registry was established in the Netherlands containing data directly sampled from electronic health record systems of physical therapists (PTs). This registry aims to evaluate the potential of patient reported outcome measures (PROMs) to develop quality indicators (QIs) in physical therapy care.

Purpose

To test to what extent the collected PROM data are reliable, valid and discriminatory between practices in measuring outcomes of patients with non-specific low back pain (NSLBP).

Methods

In this retrospective cohort study 865 PT practices with 6,560 PTs voluntarily collected PROM data of patients with NSLBP, using the Quebec Back Pain Disability Scale (QBPDS), the Numeric Pain Rating Scale (NPRS) and the Patient Specific Functioning Scale (PSFS). Reliability was determined by analysing the completeness of the dataset, the comparability by using national reference data, and through checking selection bias in the included patients. Validity was tested using the known-groups contrast between patients with (sub)acute vs. chronic NSLBP. To determine discriminative ability of outcomes between PT practices, case-mix corrected hierarchical multilevel analyses were performed.

Results

Reliability was sufficient by confirming fifteen of the sixteen hypotheses: 59% of all patients opted in for data analysis, 42% of these included patients showed repeated measurement, comparing with reference data and potential selection bias showed < 5% between group differences, while differences between (sub)acute and chronic NSLB-groups were significantly larger than 5% (less treatment sessions, lager differences in outcomes in (sub)acute NSLB patients). In addition, all nine adjusted hierarchical multilevel models confirm that the collected dataset on outcomes in PT care is able to discriminate between practices using PROM results of patients with NSLBP (ICC-scores range 0.11–0.21).

Limitations

Although we have shown the reliability, validity and discriminative ability of the dataset in the quest to develop QIs, we are aware that reducing missing values in patient records and the selective participation of PTs that belong to the innovators needs attention in the next stages of implementation to avoid bias in the results.

Conclusion

PROMs of patients with NSLBP collected in the national clinical registry of KNGF are reliable, valid and able to discriminate between primary care PT practices.

Introduction

Within healthcare, the use of health outcomes as quality indicators (QIs) to enable transparency of the service delivery is gaining momentum [15]. Although seemingly promising, there is still limited experience with the development and implementation of QIs based on health outcomes [6, 7]. Presumably, data collected in clinical registries can be used in developing such QIs, ultimately assisting in achieving more transparency of service delivery [810].

QIs are “measurement tools, screens, or flags that are used as guides to monitor, evaluate, and improve the quality of patient care, clinical support services, and organizational functions that affect patient outcomes” [11] (p 524). QIs consist of explicitly defined and measurable items referring to the structures, processes or outcomes of care [1214], and have the potential to be used to support quality improvement, accountability and to provide transparency of service delivery in healthcare [15]. A prerequisite of QIs to be of added value, is that they are defined using data, for example from KNGF’s clinical registry, that are valid, reliable and able to discriminate between groups of interest [2].

The Royal Dutch Society for Physical Therapy (KNGF) aims to develop and implement QIs using data from a clinical registry. To do so they initiated a quality program called Quality in Motion [16]. In this program a national clinical registry has been developed that enables the anonymous collection of data on: patient characteristics, structure, processes and (patient reported) health outcomes. Data are only collected after a one-off informed consent was provided by the patient and their therapist and recorded in the patient file. The data is collected directly from the electronic health record (EHR) systems used in primary care physical therapy (PT) practices [16].

Clinical practice guidelines (CPGs) often include recommendations on healthcare delivery and outcomes that can be transformed into QIs [12, 1720]. An example of such a recommendation is the use of (patient reported) outcomes that measure health-related aspects such as physical functioning or perceived pain and can be used to evaluate treatment [2126]. Since the KNGF has a long history in developing high quality CPGs [22, 27], they decided to use the recommendations in their CPGs for the selection of patient reported outcome measures (PROMs) as a basis for developing QIs. PROMs were chosen for the development of QIs, since they enable the measurement of health outcomes based on the direct perception of patients and these instruments are recommended in the high quality CPGs of KNGF.

The purpose of this study was to test to what extent health outcomes collected with PROMs in the clinical registry of KNGF are reliable, valid and able to discriminate between PT practices. These psychometric properties have been tested on data collected in patients with non-specific low back pain (NSLBP), a patient category with a large prevalence in physical therapy care. The following research questions were formulated:

  1. What is the reliability and validity of health outcomes in the clinical registry measured with PROMs?

  2. To what extent do health outcomes of physical therapy care collected in the clinical registry discriminate between practices using PROM results of patients with NSLBP?

Materials and methods

Study design and setting

This is a retrospective observational cohort study based on data collected in KNGF’s clinical registry by PT practices in Dutch primary care physical therapy. All data collected between 01-01-2013 until 28-11-2018 were used.

Participants

A total of 865 PT practices with 6,560 physical therapists (PTs) voluntarily collected data of patients with NSLBP.

Data collection

Data on all patient categories that visited the PT were collected in the registry. This was done by uploading anonymous data from patient records in the clinically used EHR-systems to KNGF’s clinical registry. To ensure uniformity of the data collection [28], all data in the registry are collected based on predefined technical specifications [16]. This procedure of data collection has shown to be feasible [16].

Data were only uploaded to the clinical registry from the EHR-systems if: 1) the patient provided their informed consent for anonymous use of the clinical data from their patient record, and 2) the therapist provided their informed consent on the use of all patient records for which they received an informed consent of the patient. For the initial reliability analysis that focussed on the completeness and comparability, all data from the clinical registry were used. For the other analyses, that specifically aimed on patients with NSLBP a selection was made and data of patients were only included if: patients visited their PT due to NSLBP, the treatment episode was finished, patients were 18 years or older, if gender and the level of chronicity of their complaints was recorded (i.e. (sub)acute vs. chronic, with (sub)acute being 0–12 weeks since onset, and chronic >12 weeks [29]), and if a PT practice had collected data from at least 30 patients with NSLBP.

Selected outcome measures

We used the PROMs that were recommended in KNGF‘s clinical practice guideline (CPG) for low back pain [25]. These PROMs are: the Quebec Back Pain Disability Scale (QBPDS), the Numeric Pain Rating Scale (NPRS) and the Patient Specific Functioning Scale (PSFS).

The QBPDS is a 20-item PROM which measures the domain physical functioning using a 6-point Likert scale, ranging from 0 “not difficult at all” to 5 “unable to do”. The overall score on the QBPDS is the sum score of the 20 items. A minimum score (0) represents “not being disabled at all” and the maximum score (100) represents “being maximally disabled” [30, 31]. The QBPDS is a feasible PROM which takes patients approximately 10 minutes to administer. The QBPDS scores limited to moderate evidence for good reliability, validity and usability [32]. In addition, it has been identified as a responsive PROM with a minimal important change (MIC) of 20 points [32, 33].

The NPRS is an instrument that assesses the domain perceived pain intensity [34]. It is a 1-item questionnaire with an 11-point scale ranging from 0 “no pain” to 10 “extreme pain” [35, 36]. The NPRS is identified to be feasible and easy to administer [35, 36]. The NPRS scores moderate to high on the psychometric properties reliability and validity [34, 37] and is responsive with a MIC of 2 points [37, 38].

The PSFS, which is equivalent to the Patient Specific Complaints (PSC) instrument [39], measures the domain physical functioning and “involves four steps, in which the patient’s main activity problems are identified, prioritized, scored and evaluated” [40] (p. 2). Each activity is scored on a 11-point scale ranging from 0 “Able to perform activity at pre-injury level” to 10 “Unable to perform activity” [41]. The PSFS is a feasible instrument with good to excellent measurement properties (reliability, validity and sensitivity to change) [38, 41]. The MIC for the PSFS was set at 2 points [42].

Sample size

In deciding if the selected data exceeds the sample size threshold, a general rule of thumb concerning the ability to discriminate outcomes between practices states that a minimum of 30 PT practices are required which should include a minimum of 30 patients each equalling 900 patients [28, 43]. The included PROMs in this study measure the domains physical functioning, and pain intensity. Perreault and Dionne (2005) estimated the agreement between patients and PTs on the domain physical functioning (ICC = 0.56) and pain intensity (ICC = 0.55) [44]. This enabled us to calculate a more specific sample size, using the equation of Twisk (2013), presented in S1 Appendix [45]. For this study this resulted in a minimum sample size of 1963 patients, treated in 66 PT practices.

Data analysis

Reliability of the data

The reliability of the included data was determined in four ways. First, by determining the completeness of the complete dataset of the clinical registry. This was done using the mean percentage of patients that were included (opted-in) in the database from the EHR-systems. Since data collection in the clinical registry is innovative and involves the anonymous processing of personal data on health status, we assumed that not all patients were willing or invited by their therapist to cooperate, which may result in selection bias if the number of patients that did not opt-in is too high. Given the early stage of implementation of the registry, we hypothesized it to be realistic to aim for a percentage of opted-in patients that lies above 50%. In addition, using the same dataset, completeness was determined by calculating the mean percentage of the (opted-in) patient records where a repeated pre- and post-treatment measurement with one of the selected PROMs was executed. Based on previous studies in the Netherlands and Israel, we hypothesized it to be realistic to set this percentage at a minimum of 40% [16, 46].

The second evaluation of reliability was estimating comparability of patient characteristics in the opted-in population in the complete clinical registry with national reference data. We compared the characteristics age and gender of all patient records of the clinical registry with a dataset that is considered to be the national reference data [47]. This analysis enabled us to check if the data in the registry is comparable to a dataset which is seen as a national benchmark. The age categories chosen for this study were aligned with those used in the national reference dataset.

The third evaluation estimated the potential selection bias in the dataset of patients with NSLBP selected from the clinical registry based on inclusion rates per practice. This analysis was executed by dividing the records of patients with NSLBP in the clinical registry into two groups: practices with low vs. practices with high inclusion rates. The two groups were created using the median percentage of patients per PT practice that are included (opted-in) in the registry, based on the patient’s informed-consent. The group with low inclusion rates was compared to the group with high inclusion rates on the characteristics age, gender and number of treatment sessions. This analysis enabled us to test again if ‘the percentage of opted-in patients’ created selection bias in the included population, which in turn would decrease reliability.

For the fourth reliability analysis, we further estimated potential selection bias in the dataset on patients with NSLBP through a different approach. The presence or absence of a repeated measurement with one of the selected PROMs (QBPDS, NPRS or PSFS) was used to create two groups. The group with a repeated measurement was compared to the group without a repeated measurement on the characteristics age, gender and number of treatment sessions. This enabled us to check if ‘the presence/absence of a repeated PROM measurement’ created selection bias in the included population, which in turn would decrease reliability.

Due to the large sample sizes it was expected that, using a t-test, the differences between the groups for the second, third and fourth reliability analyses would be statistically significant on all items. Nevertheless, we hypothesized that the analyses would not result in relevant differences between the created groups. We set an a priori threshold of >5% difference as relevant, i.e. groups were considered equal if the differences were ≤5%. Differences between the created categories were established by calculating the relative differences on the outcomes of the created groups.

Validity of the data

Regarding the validity of the collected data on patients with NSLBP the known-groups validity was determined, which is a component of construct validity frequently used to determine psychometric properties of measurement instruments [48]. With known-groups validity, the ability to distinguish or discriminate among distinct groups is defined [49]. We expected that distinct groups in the clinical registry would be present, and we defined the known-groups to be patients with chronic NSLBP versus those with (sub)acute NSLBP. Based on the findings of Costa et al. (2012) we hypothesized that patients with chronic NSLBP would need a higher number of treatment sessions and would achieve a lower change score on the three PROMs of interest (QBPDS, NPRS and PSFS) [29]. Due to the large sample sizes it was expected that, using a t-test, the differences between the groups would be statistically significant on all items, therefore we set an a priori threshold of >5% difference between the groups as being relevant. Differences between the created categories were established by calculating the relative differences on the outcomes of the created groups. For this analysis only the records of patients with NSLBP were included in which data of at least one of the selected PROMs was available at baseline (T0) and endpoint (Tend) of treatment.

The ability to discriminate between practices

To determine the ability of the collected dataset on patients with NSLBP in the clinical registry to discriminate outcomes between PT practices using the PROM results, several hierarchical multilevel analyses were performed. Initially, for each of the outcome measures -QBPDS, NPRS and PSFS- three intercept-only hierarchical multilevel models were estimated. Each model resulted in an intraclass correlation coefficient (ICC), representing the ability of the collected data to discriminate outcomes between PT practices. If an ICC is >0.10 it can be interpreted as adequate, indicating that the model is able to discriminate outcomes between PT practices [50]. The ICC values in discriminating outcomes of healthcare typically range between 0.05 and 0.20 [51]. The three different models were estimated since they represent the possible outcomes of the collected PROMs. The first model estimated the mean pre-posttreatment (T0—Tend) change scores with 95% confidence intervals (CI) of individual practices. The second model estimated the mean percentage and 95% CI of patients in which the MIC was achieved of individual practices. The third model estimated the mean post-treatment score (Tend) with 95% CI on the selected PROMs of individual practices. For each of the three models overall mean scores for all practices combined were also estimated.

Next, it was tested if stronger models were created when adjusted models were estimated, applying case-mix correction for the independent variables: age, gender, chronicity of the complaints (i.e. (sub)acute vs. chronic, with (sub)acute being 0–12 weeks since onset and chronic >12 weeks) and severity of complaints at the start of the treatment (using the results on the PROMs). These independent variables were seen as contextual variables that may influence the results on the outcome measures. As defined by Twisk (2019) [51], all such variables can be included in the model at once [51]. The ICC presents “a good gauge of whether a contextual variable has a significant effect on the outcome” [52] (p 818). The model with the highest ICC value represents the strongest model and was selected to estimate the ability of the collected data in the clinical registry to discriminate outcomes between PT practices.

Differences between groups for reliability analyses were statistically tested using independent-sampled t-tests or Fisher’s Exact test if the assumption on normality was met. IBM SPSS Statistics for Windows, version 23 (IBM Corp., Armonk, NY) was used for all analyses.

Ethical considerations

This study was conducted according to the principles of the Declaration of Helsinki (version October 2013) and in accordance with the Medical Research Involving Human Subjects Act (WMO). The study protocol was approved by the regional Medical Ethical Committee of Radboud university medical center (registration #2014/260).

Results

Overall the clinical registry contained 213,245 records of patients with NSLBP, collected from 865 PT practices, see S1 Fig for a flowchart of the patient inclusion. The mean age of the patients was 52.4 years (SD = 17.2), 55.0% of the patients was female and 77.5% of the patients had (sub)acute complaints. The registry contained 21,758, 54,904 and 73,554 patient episodes with repeated measurements of the QBPDS, NPRS and PSFS, respectively (See Table 1). See Table 2 for a summary on all executed analyses on the different psychometric properties which are explained below.

Table 1. Descriptive statistics of mean PROM-scores (unadjusted for T0 scores) at practice level.

QBPDS* NPRS# PSFS^
(score range: 0–100) (score range: 0–10) (score range: 0–10)
PT practices (N) 204 405 500
Patients (N) 21,758 54,904 73,554
Mean baseline score (SD) 40.0 (18.4) 6.3 (1.8) 6.9 (1.9)
Mean end score (SD) 13.1 (15.9) 2.2 (2.1) 1.9 (2.4)
Mean change T0-Tend (SD) -27.0 (19.8) -4.2 (2.5) -5.0 (2.7)

* QBPDS–Quebec Back Pain Disability Scale.

# NPRS–Numeric Pain Rating Scale.

^ PSFS—Patient Specific Functioning Scale.

Table 2. A summary on all executed analyses on the different psychometric properties.

Psychometric property Aspect of interest Determined by:
Reliability Completeness     • Calculating the percentage of patients that were opted-in. It was hypothesized to be realistic to aim for a percentage of opted-in patients that lies above 50%.
    • Calculating the mean percentage of the (opted-in) patient records where a repeated pre- and post-treatment measurement with one of the selected PROMs was executed. It was hypothesized to be realistic to set this percentage at a minimum of 40%.
Comparability Comparing the patient characteristics age and gender of all patient records of the clinical registry with a dataset that is considered to be the national reference data. It was hypothesized that the analyses would not result in relevant differences.
Selection bias     • Dividing the records of patients with NSLBP in the clinical registry into two groups: practices with low vs. practices with high inclusion rates. The two groups were created using the median percentage of patients per PT practice that are included (opted-in) in the registry, based on the patient’s informed-consent. The group with low inclusion rates was compared to the group with high inclusion rates on the characteristics age, gender and number of treatment sessions. It was hypothesized that the analyses would not result in relevant differences between the created groups.
    • Dividing the records of patients with NSLBP in the clinical registry into two groups based on the presence or absence of a repeated measurement with one of the selected PROMs (QBPDS, NPRS or PSFS). The groups were compared on the characteristics age, gender and number of treatment sessions. It was hypothesized that the analyses would not result in relevant differences between the created groups.
Validity Known-groups validity Determined by analysing the differences on the variables number of treatment sessions and achieved change score on the used PROMs. It was hypothesized that patients with chronic NSLBP would need a higher number of treatment sessions and would achieve a lower change score on the three PROMs of interest (QBPDS, NPRS and PSFS).
Discriminant ability Intraclass Correlation Coefficients Executing several hierarchical multilevel analyses, both intercept-only and adjusted models. If an ICC is >0.10 it can be interpreted as adequate, indicating that the model is able to discriminate outcomes between PT practices.

Reliability of the collected data

The analysis on the completeness, using the complete dataset of the clinical registry, showed that 59.2% of the patients provided permission for the use of their data, thus are opted-in. In total 41.7% of the complete dataset of the clinical registry has executed a repeated measurement with one of the selected PROMs. Both percentages exceed the benchmark that we have defined, confirming our hypotheses on completeness.

Comparing data from the complete dataset of the clinical registry with the national reference data showed a statistically non-significant difference in gender of 0.2%, and significant differences in all age groups with percentages between -2.3% and 2.5% (see Table 3). None of the comparisons exceeded the a priori defined threshold of 5% for being relevant differences, confirming our hypothesis on comparability.

Table 3. Reliability analysis on comparability.

Total clinical registry National reference data [47] Difference (%)*
Age distribution:
PT practices (N) 1,812 N/A N/A
Patients (N) 1,377,215 29,326 N/A
Patients aged 0–4 (N (%)) 9,522 (0.7) 59 (0.2) 0.5
Patients aged 5–17 (N (%)) 87,125 (6.3) 1,438 (4.9) 1.4
Patients aged 18–44 (N (%)) 404,583 (29.4) 7,8892 (6.9) 2.5
Patients aged 45–64 (N (%)) 499,483 (36.3) 10,3813 (5.4) 0.9
Patients aged 65–74 (N (%)) 210,511 (15.3) 4,8091 (6.4) 1.1
Patients aged 75–84 (N (%)) 126,351 (9.2) 3,372 (11.5) -2.3
Patients aged ≥ 85 (N (%)) 39,640 (2.9) 1,378 (4.7) 1.8
Gender distribution:
PT practices (N) 1,812 N/A N/A
Patients (N) 1,399,926 29,326 N/A
Male patients (N (%)) 562,632 (40.2) 11,730 (40.0) 0.2
Female patients (N (%)) 837,294 (59.8) 17,596 (60.0) -0.2

All data of the KNGF’s clinical registry compared to the national reference data [47].

* Percentual differences are calculated by subtracting the percentages of the National reference data from the Total clinical registry data.

†p ≤ 0.001 (t-test).

N/A: Not applicable.

Estimating the potential selection bias in the included patients with NSLBP, the analysis comparing practices with high versus low inclusion rates of patients, resulted in non-significant differences of -0.2%, 1.1% and -2.8% on age, gender and number of treatment sessions, respectively (see Table 4). The analysis comparing presence or absence of a repeated measurement with one of the selected PROMs, using the same dataset, resulted in statistically significant differences, of 2.7%, -0.2% and 35.0% on age, gender and number of treatment sessions, respectively (see Table 5). Five of the six analyses confirmed our hypotheses on selection bias since they did not exceed the a priori defined threshold of 5%.

Table 4. Reliability analysis on selection bias using the NSLBP sample (N = 213,245 patients from 865 PT practices) of the total clinical registry: A within groups comparison on descriptive statistics using the median opt-in to divide the NSLBP sample into two groups.

NSLBP-sample of the total clinical registry scoring below median^ opt-in NSLBP-sample of total clinical registry scoring above median^ opt-in Difference (%)*
Mean age patients (SD) 52.6 (4.6) 52.7 (4.9) -0.2
Percentage of female patients (SD) 55.5 (8.1) 54.9 (7.7) 1.1
Mean number of treatment sessions (SD) 6.9 (2.8) 7.1 (2.6) -2.8

^ median % patients included = 68.3.

* Percentual differences are calculated using the following formula: Difference (%) = ((NSLBP-sample of the total clinical registry scoring below median^ opt-in / NSLBP-sample of total clinical registry scoring above median^ opt-in) * 100) -100.

Table 5. Reliability analysis on selection bias using the NSLBP sample of the total clinical registry: A within groups comparison using the availability of a repeated measurement with a PROM (NPRS/ QBPDS/ PSFS) to divide the NSLBP sample in two groups.

With pre- and post-test PROM use Without pre- and post-test PROM use Difference (%)*
PT practices (N) 775 865 N/A
Patients (N) 88,852 124,393 N/A
Mean age patients (SD) 53.2 (17.1) 51.8 (17.2) 2.7
Percentage of female patients (SD) 55.6 (5.7) 55.7 (6.1) -0.2
Mean number of treatment sessions (SD) 8.1 (8.9) 6.0 (8.3) 35.0

* Percentual differences are calculated using the following formula: Difference (%) = ((With pre- and post-test PROM use / Without pre- and post-test PROM use) * 100) -100.

†p ≤ 0.001 (T-test).

N/A: Not applicable.

Known-groups validity of the collected data

Patients with (sub)acute NSLBP needed 25.7% less treatment sessions than patients with chronic NSLBP. The three other analyses showed that patients with (sub)acute NSLBP compared to patients with chronic NSLBP achieved a 40.1%, 29.4% and 23.8% higher change score on the QBPDS, NPRS and the PSFS, respectively. The results of these four analyses for the known-groups validity all exceeded the set threshold of 5%, confirming our hypothesis on the known-groups validity of the data (see Table 6).

Table 6. Validity analysis on the NSLBP sample of the total clinical registry, using the known-groups validity to divide the NSLBP-sample in two groups.

Patients with (sub)acute NSLBP Patients with chronic NSLBP Difference (%) ¤
Mean (SD) PT practices (N) Patients (N) Mean (SD) PT practices (N) Patients (N)
Number of treatment sessions 7.5 (7.7) 526 65,284 10.1 (11.3) 472 18,658 -25.7
Change score PROM T0 versus Tend:
• QBPDS* -28.6 (19.7) 204 17,399 -20.3 (19.0) 186 4,359 40,1
• NPRS# -4.4 (2.5) 405 43,467 -3.4 (2.6) 350 11,447 29.4
• PSFS^ -5.2 (2.7) 500 57,284 -4.2 (2.9) 439 16,270 23.8

¤ Percentual differences are calculated using the following formula: Difference (%) = ((Patients with (sub)acute NSLBP Mean (SD) / Patients with chronic NSLBP Mean (SD)) * 100) -100

†p ≤ 0.001 (t-test).

* QBPDS–Quebec Back Pain Disability Scale.

# NPRS–Numeric Pain Rating Scale.

^ PSFS—Patient Specific Functioning Scale.

The ability to discriminate between practices based on the collected data

Seven of the nine ICC-scores obtained from the intercept-only hierarchical multilevel analyses exceed the a priori formulated threshold of 0.10 (see Table 7). In the adjusted-models all nine ICC-scores exceeded this threshold with scores ranging from 0.11–0.21. These results indicate that the adjusted-models are able to discriminate outcomes between PT practices, confirming the discriminative ability of the data [50].

Table 7. The intraclass correlation coefficients for the intercept-only model and adjusted model on the mean change-score, the mean percentage of MIC achieved-score and the mean end score of all selected PROMs.

ICC intercept-only model ICC adjusted-model
Mean change score:
NPRS# 0.10 0.16
QBPDS* 0.12 0.21
PSFS^ 0.11 0.17
Mean percentage MIC achieved score:
NPRS# 0.11 0.14
QBPDS* 0.12 0.19
PSFS^ 0.12 0.14
Mean end score:
NPRS# 0.12 0.13
QBPDS* 0.10 0.11
PSFS^ 0.12 0.14

ICC–Intraclass Correlation Coefficients.

# NPRS–Numeric Pain Rating Scale.

* QBPDS–Quebec Back Pain Disability Scale.

^ PSFS—Patient Specific Functioning Scale.

Discussion

Overall the results of this study show that the analyses regarding the reliability of the data in the clinical registry were in line with our a priori formulated hypotheses: the data was sufficiently complete, comparable and we did not identify selection bias in the patients that were included in the registry. All analyses on the known-groups validity met the a priori formulated hypotheses. Based on these results it can be concluded that both reliability and known-groups validity are confirmed. In addition, the hierarchical multilevel analyses confirm that the collected dataset on outcomes in PT care is able to discriminate between practices using PROM results of patients with NSLBP.

Only one of the six analyses aiming to check for selection bias in the included patients, as part of the reliability analyses, did not meet the a priori formulated hypothesis. This analysis represented the number of treatment sessions required in patients with a repeated PROM measurement versus those without. The results showed that patients who required fewer treatments more often did not complete a repeated PROM measurement. This might be explained by the fact that PT’s and/or patient’s belief that the completions of a repeated PROM measurement have no added value for this patient category with a short treatment period and fast recovery.

To the best of our knowledge, within the field of physical therapy similar studies, which execute a so-called practice test, aiming to investigate the reliability, validity and discriminative ability of the collected data in a clinical registry, have not been published. In several countries and healthcare settings similar initiatives started the collection of PROMs scores in clinical registries, as a prerequisite for the development of QIs based on clinical data. A first example of such registry is initiated by the National Health Service in the United Kingdom, which introduced their national PROMs programme in 2009. A second example is initiated in 2010 by the Dutch Institute for Clinical Auditing (DICA). DICA manages 22 clinical registries, belonging to different scientific associations of medical specialists (e.g. neurologists & oncologists) [53]. Despite this experience in building clinical registries, several publications suggest that the integrated implementation of PROMs, using them to stimulate shared decision-making on clinical level and as performance information on managerial level has not yet been successfully achieved [8, 54].

In the quest for developing QIs using data from clinical registries, the execution of a practice test to evaluate the psychometric properties of the collected data is crucial. There are a limited number of studies in other fields than PT that have conducted such a practice test. Examples of such studies are analysed in the review on the validity of QI’s on the readmission rate of Fischer et al. (2012). This study showed that only a small proportion 21 of the 486 included papers test the actual validity of the data in some sort of practice test [55]. In addition, a systematic review of Langendam et al. (2020) showed that performing a practice test to validate the formulated QI which are based on recommendations of CPGs is relatively rare [20].

There are some publications providing a framework on the development of QIs [18, 56]. These frameworks also emphasize the importance of the execution of a practice test, since it enables formulating a benchmark for a QI that meets all psychometric criteria [2, 28, 56]. Unfortunately, several studies state that there is no standard definition of what a practice test has to contain [18, 20]. Often only the Delphi methodology is used to reach consensus in formulating QIs and practice tests are not executed [17, 19, 42, 57].

In short, there is no guidance available for applying a practice test as we have done in this study. The results of this study are therefore a good starting point for the development of standards for a practice test to be carried out in the development of benchmarks for QIs based on data from a clinical registry that meet all relevant psychometric properties.

Strengths and limitations

The strength of this paper is the large number of PTs that voluntarily participated and collected data on a very large number of participating patients.

Nevertheless, this paper is subject to several limitations. Despite the fact that the infrastructure of the clinical registry has been successfully built up, we identified that input fields from the patient files in the EHR-systems are still registered as missing values, and thus could not be included in this study. These missing values sometimes occurred due to technical omissions (e.g. there are several QBPDS questionnaires defined in the EHR-software but only one of the questionnaires is eligible for extraction to the registry), but also due the submission of incompletely registered patient files from the EHR-system to the national registry. In the further implementation of the clinical registry, we make a continuous effort in trying to decrease the number of missing values. A specific challenge is the large number of different EHR-suppliers that is involved. We have improved the collaboration with the different EHR-system suppliers, which assists in overcoming the technical omissions. In addition, we have executed different projects trying to influence the behaviour of PTs in completing all fields of a patient file before uploading. Examples of such projects are: the organization of several regional symposia to teach PTs on how to use the registry and the development of an online (free for use) feedback dashboard. Such implementation activities are important in the further exploitation of a clinical registry. After all, having a registry with complete patient files is crucial in the development of valid benchmarks for QIs. A second limitation is the chance of selection bias based on the PT practices that participate in the data collection. This is likely to have happened since -in this early stage of implementation- the current users of the clinical registry, the PTs who voluntarily provided all data, are mainly the innovators, early adopters and early majority [58]. These participants are probably PTs who favour the use of PROMs. A third limitation might be the fact that we have used somewhat older data from the registry. Despite this, we assume that the data are still relevant as a good representation of the clinical practice and that they shed light on the potential that a clinical registry has in formulating QIs and their benchmarks. Moreover, the outcomes can be used as baseline measurement to evaluate changes over time.

Implications for future research and clinical practice

Future research should focus on several topics. First, the further development of a standardised practice test, which is an important step before defining QIs and their benchmarks. Such a test should finally include all psychometric properties that need to be tested in defining QIs. These aspects are: reliability, validity, discriminative ability, responsiveness to change, feasibility and usability [2, 28]. The development of such a practice test will likely help with the implementation of QIs in clinical practice and will minimize the resistance in the use of QI and their benchmark. Second, research should focus on the development of actual benchmark based on the collected data thus far. This means that an extensive developmental process must be completed before a QI, including a psychometrically sound benchmark is finished. Given the investment needed it does not seem feasible to develop QIs for all conditions in the field of PT. Therefore it is relevant to investigate what solution can be found for this problem. A possible solution might be the development of several generic QIs for the field of PT and only develop specific QIs for conditions that are seen on a very regular basis by all PTs, as is the case in patients with NSLBP.

In addition, from a clinical practice perspective, efforts should be made to assist PTs in completing their patient files (e.g. by using modern technology such as applications for mobile devices which enable patients to fill in PROMs on their mobile phones) and to increase the number of PTs who will provide their data to the clinical registry. Finally, continuous investments must be made in improving the infrastructure of the clinical registry, on the one hand to detect errors (e.g. missing values) and on the other to improve efficiency and ease of use.

Conclusions

This study showed that the health outcomes of patients with NSLBP collected with PROMs in the national clinical registry of KNGF are reliable, valid and are able to discriminate between primary care PT practices.

Supporting information

S1 Appendix. The calculation used to define a more specific sample size, using the equation of Twisk et al. (2013).

(DOCX)

S1 Fig. A flowchart of the patient inclusion.

(DOCX)

S1 Syntax. 20180309_cleaning_definitief_GUUS_JC.sps.

(SPS)

S2 Syntax. 20201603 descriptives and provisionary analysis and QI analysis.sps.

(SPS)

S1 Dataset. 20180116_Totaalbestand_aanlever.sav.

(ZIP)

S2 Dataset. werkbestand_20180308_DCSPH.sav.

(ZIP)

S3 Dataset. werkbestand_20180308_DCSPH_leeftijd.sav.

(ZIP)

S4 Dataset. werkbestand_20180308_DCSPH_leeftijd_geslacht.sav.

(ZIP)

S5 Dataset. werkbestand_20180308_DCSPH_leeftijd_geslacht_afgesloten.sav.

(ZIP)

S6 Dataset. werkbestand_20180308_DCSPH_leeftijd_geslacht_afgesloten_duurklachten.sav.

(ZIP)

S7 Dataset. 20180314_multilevel.sav.

(ZIP)

S8 Dataset. One sample T-test Nivel data vs LDF data.sav.

(ZIP)

Acknowledgments

We would like to acknowledge all patients, physical therapists and practices who voluntarily collected the valuable data into KNGF’s clinical registry over the past years. In addition, we would like to thank our research colleagues Marjo Maas, Janine Liefers and Annick Bakker-Jacobs for all the time they have invested over the past years in training PTs on how to work with the registry and in their role in analysing the data which results in valuable information for the academic, the clinicians and ultimately for the patient.

Data Availability

We are happy to be able to share the anonymized dataset with PLOS ONE. This enables us to comply with the important academic FAIR principles. We have added all necessary datasets and syntaxes as Supporting Information files. In uploading the datafiles (SPSS) we had to compress (.ZIP) the files.

Funding Statement

The Quality in Motion program was funded by the association KNGF (Koninklijk Nederlands Genootschap voor Fysiotherapie). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Karthaus EG, Vahl A, Kuhrij LS, Elsman BHP, Geelkerken RH, Wouters MWJM, et al. The Dutch Audit of Carotid Interventions: Transparency in Quality of Carotid Endarterectomy in Symptomatic Patients in the Netherlands. Eur J Vasc Endovasc Surg. 2018;56(4):476–485. doi: 10.1016/j.ejvs.2018.05.030 [DOI] [PubMed] [Google Scholar]
  • 2.Fischer C. Quality Indicators for Hospital Care. (Theunisse I, ed.). Ede: GVO drukkers & vormgevers B.V; 2015. [Google Scholar]
  • 3.Brand CA, Martin-Khan M, Wright O, Jones RN, Morris JN, Travers CM, et al. Development of quality indicators for monitoring outcomes of frail elderly hospitalised in acute care health settings: Study Protocol. BMC Health Serv Res. 2011;11(1):281. doi: 10.1186/1472-6963-11-281 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Kuhrij LS, Wouters MWJM, van den Berg-Vos RM, de Leeuw FE, Nederkoorn PJ. The Dutch Acute Stroke Audit: Benchmarking acute stroke care in the Netherlands. Eur Stroke J. 2018;3(4):361–368. doi: 10.1177/2396987318787695 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Petersson IF, Strömbeck B, Andersen L, Cimmino M, Greiff R, Loza E, et al. Development of healthcare quality indicators for rheumatoid arthritis in Europe: The eumusc.net project. Ann Rheum Dis. 2014;73(5):906–908. doi: 10.1136/annrheumdis-2013-203765 [DOI] [PubMed] [Google Scholar]
  • 6.Govaert JA, Dijk WA Van, Fiocco M, Scheffer AC. Nationwide Outcomes Measurement in Colorectal Cancer Surgery: Improving Quality and Reducing Costs. J Am Coll Surg. 2016;222(1):19–29.e2. doi: 10.1016/j.jamcollsurg.2015.09.020 [DOI] [PubMed] [Google Scholar]
  • 7.Overveld van L. Quality indicators of integrated care for patients with head and neck cancer. Clin Otolaryngol. 2017;42:322–329. doi: 10.1111/coa.12724 [DOI] [PubMed] [Google Scholar]
  • 8.van der Wees P, Nijhuis-van der Sanden M, Ayanian J, Black N, Westert G, Schneider E. Integrating the Use of Patient-Reported Outcomes for Both Clinical Practice and Performance Measurement: Views of Experts from 3 Countries. Milbank Q. 2014;92(4):754–775. doi: 10.1111/1468-0009.12091 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Black N. Patient reported outcome measures could help transform healthcare. Bmj. 2013;346(January):2–5. doi: 10.1136/bmj.f167 [DOI] [PubMed] [Google Scholar]
  • 10.Wu AW, Kharrazi H, Boulware LE, Snyder CF. Measure once, cut twice—Adding patient-reported outcome measures to the electronic health record for comparative effectiveness research. J Clin Epidemiol. 2013;66(SUPPL.8):S12–S20. doi: 10.1016/j.jclinepi.2013.04.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Mainz J. Defining and classifying clinical indicators for quality improvement. Int J Qual Heal Care. 2003;15(6):523–530. doi: 10.1093/intqhc/mzg081 [DOI] [PubMed] [Google Scholar]
  • 12.Westby MD, Klemm A, Li LC, Jones CA. Emerging Role of Quality Indicators in Physical Therapist Practice and Health Service Delivery. Phys Ther. 2016;96(1):90–100. doi: 10.2522/ptj.20150106 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.McGlynn EA, Asch SM. Developing a clinical performance measure. Am J Prev Med. 1998;14(3 SUPPL.):14–21. doi: 10.1016/s0749-3797(97)00032-9 [DOI] [PubMed] [Google Scholar]
  • 14.Centers for Medicare and Medicaid Services. https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/QualityMeasures/index.html. Published 2019. Accessed January 16th 2021
  • 15.Mainz JAN. Developing evidence-based clinical indicators: a state of the art methods primer. Int J Qual Heal Care. 2003;15(supplement I):5–12. doi: 10.1093/intqhc/mzg084 [DOI] [PubMed] [Google Scholar]
  • 16.Meerhoff G, van Dulmen S, Maas M, Heijblom K, Nijhuis-van der Sanden M, van der Wees P. Development and Evaluation of an Implementation Strategy for Collecting Data in a National Registry and the Use of Patient Reported Outcome Measure (PROMs) in Physical Therapist Practice: Quality Improvement Study. Phys Ther J Am Phys Ther Assoc. 2017;97(8):837–851. doi: 10.1093/ptj/pzx051 [DOI] [PubMed] [Google Scholar]
  • 17.Kötter T, Blozik E, Scherer M. Methods for the guideline-based development of quality indicators-a systematic review. Implement Sci. 2012;7(1):21. doi: 10.1186/1748-5908-7-21 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Nothacker M, Stokes T, Shaw B, Lindsay P, Sipilä R, Follmann M, et al. Reporting standards for guideline-based performance measures. Implement Sci. 2016;11(6):1–11. doi: 10.1186/s13012-015-0369-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Peter WF, Hurkmans EJ, van der Wees PJ, Hendriks EJM, van Bodegom-Vos L, Vliet Vlieland TPM. Healthcare Quality Indicators for Physiotherapy Management in Hip and Knee Osteoarthritis and Rheumatoid Arthritis: A Delphi Study. Musculoskeletal Care. 2016;14(4):219–232. doi: 10.1002/msc.1133 [DOI] [PubMed] [Google Scholar]
  • 20.Langendam MW, Piggott T, Nothacker M, Agarwal A, Armstrong D, Baldeh T, et al. Approaches of integrating the development of guidelines and quality indicators: a systematic review. 2020:1–11. doi: 10.1186/s12913-020-05665-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.van Doormaal MCM, Meerhoff GA, Vliet Vlieland TPM, Peter WF.clinical practice guideline for physical therapy in patients with hip or knee osteoarthritis. Musculoskeletal Care. 2020;18(4):575–595. doi: 10.1002/msc.1492 [DOI] [PubMed] [Google Scholar]
  • 22.Bier J, Scolten-Peeters W, Staal J, Pool J, van Tulder M, Beekman E, et al. Clinical practice guideline for physical therapy assessment and treatment in patients with nonspecific neck pain. Phys Ther J Am Phys Ther Assoc. 2018;98(3):162–171. doi: 10.1089/tmj.2011.0276 [DOI] [PubMed] [Google Scholar]
  • 23.Delitto A, George SZ, Dillen LVAN, Whitman JM, Sowa G, Shekelle P, et al. Clinical Practice Guidelines Linked to the International Classification of Functioning, Disability, and Health from the Orthopaedic Section. J Orthop Sport Phys Ther. 2012;42(4):A1–A57. doi: 10.2519/jospt.2012.42.4.A1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Heemskerk M, Staal J, Bierma-Zeinstra S, de Haan G, Hagenaars L, Lanser K, et al. KNGF guideline Complaints of Arm Neck and Shoulder. Ned Tijdschr voor Fysiother. 2010;120(1 (supplement)). [Google Scholar]
  • 25.Staal J, Hendriks E, Heijmans M, Kiers H, Lutgers-Boomsma AM. Rutten G, et al. KNGF guideline Low Back Pain. Ned Tijdschr voor Fysiother. 2013;123(1 (supplement)). [Google Scholar]
  • 26.Berwick D, Black N, Cullen D, Deerberg-Wittram J, Degos L, Diverty B, et al. Recommendations to OECD Ministers of Health from the High Level Reflection Group on the Future of Healh Statistics.; 2017. https://www.oecd.org/els/health-systems/Recommendations-from-high-level-reflection-group-on-the-future-of-health-statistics.pdf. Accessed January 16th 2021. [Google Scholar]
  • 27.Hendriks H, Bekkering G, van Ettekoven H, Brandsma J, van der Wees P, de Bie R. Development and Implementation of National Practice Guidelines: A Prospect for Continuous Quality Improvement in Physiotherapy. Physiotherapy. 2000;86(10):535–547. [Google Scholar]
  • 28.van Kessel P, Triemstra M, de Boer D. Handreiking Voor Het Meten van Kwaliteit van Zorg Met Patient Reported Outcome Measures.; 2014. https://www.nivel.nl/nl/publicatie/handreiking-voor-het-meten-van-kwaliteit-van-zorg-met-patient-reported-outcome-measures. Accessed January 16th 2021. [Google Scholar]
  • 29.Costa L da CM, Maher CG, Hancock MJ, McAuley JH, Herbert RD, Costa LOP. The prognosis of acute and persistent low-back pain: a meta-analysis. Can Med Assoc J. 2012;184(11):1229–1230. doi: 10.1503/cmaj.120627 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Kopec J, Esdaile J, Abrahamowicz M, Abenhaim L, Wood-Dauphinee S, Lamping D, et al. The Quebec Back Pain Disability Scale Measurement properties. Spine (Phila Pa 1976). 1995;20(3):341–352. doi: 10.1097/00007632-199502000-00016 [DOI] [PubMed] [Google Scholar]
  • 31.Kopec J, Esdaile J, Abrahamowicz M, Abenhaim L, Wood-Dauphinee S, Lamping D, et al. The quebec back pain disability scale measurement properties. J Clin Epidemiol. 1996;49(2):151–161. doi: 10.1016/0895-4356(96)00526-4 [DOI] [PubMed] [Google Scholar]
  • 32.Speksnijder C, Koppenaal T, Knottnerus J, Spigt M, Staal J, Terwee C. Measurement Properties of the Quebec Back Pain Disability Scale in Patients With Nonspecific Low Back Pain: Systematic Review. Phys Ther J Am Phys Ther Assoc. 2016;96(11):1816–1831. doi: 10.2522/ptj.20140478 [DOI] [PubMed] [Google Scholar]
  • 33.Ostelo R, Deyo R, Strattford P. Interpreting change scores in pain and functional status in low back pain. Spine (Phila Pa 1976). 2008;33(1):90–94. http://scholar.google.com/scholar?hl=en&btnG=Search&q=intitle:Interpreting+Change+Scores+for+Pain+and+Functional+Status+in+Low+Back+Pain#5. Accessed January 16th 2021. [DOI] [PubMed] [Google Scholar]
  • 34.Kahl C, Cleland J. Visual analogue scale, numeric pain rating scale and the McGill pain Questionnaire: an overview of psychometric properties. Phys Ther Rev. 2005;10(2):123–128. doi: 10.1179/108331905X55776 [DOI] [Google Scholar]
  • 35.Williams A, Davies H, Chadury Y. Simple pain rating scales hide complex idiosyncratic meanings. Pain. 2000;85(3):457–463. doi: 10.1016/S0304-3959(99)00299-7 [DOI] [PubMed] [Google Scholar]
  • 36.Jensen M, Karoly P, Braver S. The measurement of clinical pain intensity: a comparison of six methods. Pain. 1986;27(1):117–126. doi: 10.1016/0304-3959(86)90228-9 [DOI] [PubMed] [Google Scholar]
  • 37.Childs JD, Piva SR, Fritz JM. Responsiveness of the numeric pain rating scale in patients with low back pain. Spine (Phila Pa 1976). 2005;30(11):1331–1334. doi: 10.1097/01.brs.0000164099.92112.29 [DOI] [PubMed] [Google Scholar]
  • 38.Pengel L, Refshauge K, Haher C. Responsiveness of pain, disability, and physical impairment outcomes in patients with low back pain. Spine (Phila Pa 1976). 2004;29(8):879–883. doi: 10.1097/00007632-200404150-00011 [DOI] [PubMed] [Google Scholar]
  • 39.Stratford P, Gill C, Westaway M, Binkley J. Assessing disability and change on individual patients: a report of a patient specific measure. Physiother Canada. 1995;47(4):258–263. [Google Scholar]
  • 40.Stevens A, Köke A, Van Der Weijden T, Beurskens A. Ready for goal setting? Process evaluation of a patient-specific goal-setting method in physiotherapy. BMC Health Serv Res. 2017;17(1):1–10. doi: 10.1186/s12913-016-1943-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Westaway MD, Stratford PW, Binkley JM. The Patient-Specific Functional Scale: Validation of’lts Use in Persons With Neck Dysfunction. J Orthop Sport Phys Ther. 1998;27(5):331–338. doi: 10.2519/jospt.1998.27.5.331 [DOI] [PubMed] [Google Scholar]
  • 42.Oostendorp RA, Rutten GM, Dommerholt J, Nijhuis-van der Sanden M, Harting J. Guideline-based development and practice test of quality indicators for physiotherapy care in patients with neck pain. J Eval Clin Pract. 2013;19:1044–1053. doi: 10.1111/jep.12025 [DOI] [PubMed] [Google Scholar]
  • 43.Maas CJM, Hox JJ. Robustness issues in multilevel regression analysis. Stat Neerl. 2004;58(2):127–137. [Google Scholar]
  • 44.Perreault K, Dionne CE. Patient-physiotherapist agreement in low back pain. J Pain. 2005;6(12):817–828. doi: 10.1016/j.jpain.2005.07.008 [DOI] [PubMed] [Google Scholar]
  • 45.Twisk J. Applied Multilevel Analysis. 6th ed. Cambridge: Cambridge University Press; 2013. [Google Scholar]
  • 46.Deutscher D, Horn SD, Dickstein R, Hart DL, Smout RJ, Gutvirtz M, et al. Associations Between Treatment Processes, Patient Characteristics, and Outcomes in Outpatient Physical Therapy Practice. Arch Phys Med Rehabil. 2009;90(8):1349–1363. doi: 10.1016/j.apmr.2009.02.005 [DOI] [PubMed] [Google Scholar]
  • 47.van den Dool J, Schermer T. Zorg Door de Fysiotherapeut Jaarcijfers 2018 En Trendcijfers 2014–2018.; 2019. https://nivel.nl/sites/default/files/bestanden/1003595.pdf. Accessed January 16th 2021. [Google Scholar]
  • 48.Rodrigues IB, Adachi JD, Beattie KA, Lau A, MacDermid JC. Determining known-group validity and test-retest reliability in the PEQ (personalized exercise questionnaire). BMC Musculoskelet Disord. 2019;20(1):1–10. doi: 10.1186/s12891-018-2378-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.McConnell S, Kolopack P, Davis AM. The Western Ontario and McMaster Universities Osteoarthritis Index (WOMAC): a review of its utility and measurement properties. Arthritis Rheum. 2001;45(5):453–461. doi: [DOI] [PubMed] [Google Scholar]
  • 50.Boer de D, Triemstra M, Koopman L, Zuidgeest M. Tool Analyses: Voor Het Vergelijken van Patiëntervaringen Tussen Zorgaanbieders | Versie 1.0 2016. https://www.nivel.nl/nl/publicatie/tool-analyses-voor-het-vergelijken-van-patientervaringen-tussen-zorgaanbieders-versie-10. Accessed January 16th 2021. [Google Scholar]
  • 51.Twisk JWR. Applied Mixed Model Analysis: A Practical Guide.; 2019. doi: 10.1017/9781108635660 [DOI] [Google Scholar]
  • 52.Field A. Discovering Statistics Using IBM SPSS Statistics. 4th ed. Sage Publications Ltd.; 2013. [Google Scholar]
  • 53.https://dica.nl/. https://www.dica.nl. Published 2020. Accessed November 8, 2020.
  • 54.Greenhalgh J, Dalkin S, Gooding K, et al. Functionality and Feedback: A Realist Synthesis of the Collation, Interpretation and Utilisation of Patient-Reported Outcome Measures Data to Improve Patient Care. Vol 5.; 2017. doi: 10.3310/hsdr05020 [DOI] [PubMed] [Google Scholar]
  • 55.Fischer C, Anema HA, Klazinga NS. The validity of indicators for assessing quality of care: A review of the European literature on hospital readmission rate. Eur J Public Health. 2012;22(4):484–491. doi: 10.1093/eurpub/ckr165 [DOI] [PubMed] [Google Scholar]
  • 56.van der Wees PJ, Verkerk EW, Verbiest MEA, Zuidgeest M, Bakker C, Braspenning J, et al. Development of a framework with tools to support the selection and implementation of patient-reported outcome measures. J Patient-Reported Outcomes. 2019;3(1). doi: 10.1186/s41687-019-0171-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Boulkedid R, Abdoul H, Loustau M, Sibony O, Alberti C. Using and reporting the Delphi method for selecting healthcare quality indicators: A systematic review. PLoS One. 2011;6(6). doi: 10.1371/journal.pone.0020476 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Rogers E. Diffusion of Innovations Fifth Edition. New York: Free Press; 2003. [Google Scholar]

Decision Letter 0

Alison Rushton

19 Mar 2021

PONE-D-21-03296

Reliability, validity and discriminability of patient reported outcomes for non-specific low back pain in a nationwide physical therapy registry: a retrospective observational cohort study

PLOS ONE

Dear Dr. Meerhoff,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

Please address the points raised by the two reviewers.

==============================

Please submit your revised manuscript by May 03 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Alison Rushton

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

3.  We note that you have indicated that data from this study are available upon request. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For more information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions.

In your revised cover letter, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially sensitive information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings as either Supporting Information files or to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories.

We will update your Data Availability statement on your behalf to reflect the information you provide.

4. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: 

Congratulations to the authors for such a good work. I would like to share some comments/seek for some clarification.

1. Suggest authors to expand on the importance and the use of quality indicators (QIs) in health services. It is unclear in the introduction why authors want to identify if the PROMs can discriminate PT practices.

2. Table III - Suggest to have the total N in the top row and present the data in the format of N(%). Suggest to use age categories of similar gap, if not, please justify in methods section.

3. Table IV, V, VI - how does the difference being derived? For example, mean age difference has a difference in %. Suggest author to include the explanation in "Methods" section.

4. Kindly specify test used to obtain p-values for Table III - VI. Is it t-test or Fisher's exact test?

5. Please provide the number of PT practices that were included in the analysis. Authors only described the number of patients .

Reviewer #2: 

Comments to the authors

Manuscript title:

Reliability, validity and discriminability of patient reported outcomes for non-specific low back pain in a nationwide physical therapy registry: a retrospective observational cohort study

Thank you for inviting me to review the manuscript which aimed at testing to what extent health outcomes in patients with non-specific low back pain (NSLBP) collected with PROMs in the national clinical registry of the Royal Dutch Society for Physical Therapy (KNGF) are reliable, valid and able to discriminate between physical therapy practices.

The paper provides background information about successful involvement of patients and physical therapists in a clinical registry in The Netherlands. The authors included data from an impressive number of physical therapy (PT) practices (n=865) and PTs (n=6,560) of patients with NSLBP over a long period of time, between 2013 and 2018. The comprehensive description of included outcome measures in addition to the methods used to answer the research questions were helpful. Findings were described in detail and discussed sufficiently.

This manuscript addresses an important issue for clinicians, researchers and other stakeholders as it demonstrates the importance and value of clinical registries also for formulating appropriate QI's and provides a good example of a practice test to assess the psychometric properties of the collected data. Therefore, it should be of interest to a wide audience.

I have some minor suggestions that could improve the manuscript further.

Abstract: The abstract summarizes the study very well. Please just correct “in this next stages of implementation”.

Introduction: The introduction is comprehensive and informs the reader about the importance of clinical registries also for the development of quality indicators. It would be interested how people consented to participate in the registry and if they have to consent only once or multiple times.

Method: described comprehensively. In this manuscript, acute vs. chronic NSLBP is defined with acute being 0-12 weeks since onset, and chronic >12 weeks. However, definitions of acute and chronic LBP are different, for example in a review provided by Chou et al. where acute NSLBP is defined <4 weeks' duration) while chronic/subacute NSLBP >4 weeks' duration. Please provide a reason for your definition and a reference.

Chou, R., & Huffman, L. H. (2007). Nonpharmacologic therapies for acute and chronic low back pain: a review of the evidence for an American Pain Society/American College of Physicians clinical practice guideline. Annals of internal medicine, 147(7), 492-504.

Results: are very well described. The tables provided help to overlook the results, but please look at the formatting of the tables, e.g. upper and lower lines in Table 1 missing. Table III By “total clinical registry” you mean all data of the KNGF’s clinical registry?

Discussion: very well written. I liked the idea of “implementation activities” to facilitate the use of the registry.

Line 295 “that the completions of a repeated PROM measurement has no added value for this patient category” should be have…

Further suggestions: Look for double spaces and commas throughout the text (e.g. after “In addition”, “Nevertheless”).

I wish the authors the best with their continued work in this interesting and promising area.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Jun 3;16(6):e0251892. doi: 10.1371/journal.pone.0251892.r002

Author response to Decision Letter 0


28 Apr 2021

PLOS ONE

Subject: Submission of revised manuscript

Utrecht, April 28th 2021

Dear Dr. Emily Chenette,

Thank you for the invitation to revise and resubmit our paper for publication in PLOS ONE. Please find enclosed our revised manuscript entitled: ‘Reliability, validity and discriminability of patient reported outcomes for non-specific low back pain in a nationwide physical therapy registry: a retrospective observational cohort study’.

The feedback of the reviewers was very useful. We have replied to the feedback in a point-by-point response letter including the journal requirements and the feedback from the reviewers, followed by the responses of the authors (see the Appendix). In the manuscript itself, all text revisions are visible using the track changes function of Microsoft Word (using the ‘All Markup’ mode, to keep the line numbers in correspondence with my rebuttal letter).

We hope this revision made our manuscript eligible for publication in PLOS ONE and meets your expectations.

We look forward to your kind consideration of our revised manuscript.

Yours sincerely,

Guus A. Meerhoff, MSc. (corresponding author)

Radboud university medical center, 114 IQ healthcare,

PO Box, 9101

6500 HB Nijmegen, The Netherlands

+ 31 6 48421927 (phone)

Appendix: Point by point reply to the stated journal requirements and the feedback of the reviewers

Journal requirements

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

Author’s reply:

We have changed the style of our paper to align it with PLOS ONE’s style requirements.

2. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Author’s reply:

We have checked the reference list and added references number 11-15 based on a text passage that was added in the Introduction. See also comment 1 of reviewer 1. [Line 500-511]

11. Mainz J. Defining and classifying clinical indicators for quality improvement. Int J Qual Heal Care. 2003;15(6):523-530. doi:10.1093/intqhc/mzg081

12. Westby MD, Klemm A, Li LC, Jones CA. Emerging Role of Quality Indicators in Physical Therapist Practice and Health Service Delivery. Phys Ther. 2016;96(1):90-100. doi:10.2522/ptj.20150106

13. McGlynn EA, Asch SM. Developing a clinical performance measure. Am J Prev Med. 1998;14(3 SUPPL.):14-21. doi:10.1016/S0749-3797(97)00032-9

14. Centers for Medicare and Medicaid Services. https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/QualityMeasures/index.html. Published 2019. Accessed January 16th 2021

15. Mainz JAN. Developing evidence-based clinical indicators : a state of the art methods primer. Int J Qual Heal Care. 2003;15(supplement I):5-12.

Furthermore we aligned the references with the style of PLOS ONE’s requirements (e.g. stated as a max of six authors per reference). [Line 467-642]

3. We note that you have indicated that data from this study are available upon request. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For more information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions.

In your revised cover letter, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially sensitive information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings as either Supporting Information files or to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories.

We will update your Data Availability statement on your behalf to reflect the information you provide.

Author’s reply:

We are happy to be able to share the anonymized dataset with PLOS ONE. This enables us to comply with the important academic FAIR principles. We have added all necessary datasets and syntaxes as Supporting Information files. [Line 652-680]

4. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

Author’s reply:

When aligning the paper with PLOS ONE’s style requirements we have inserted the Supporting Information including its captions in the correct format.

Comments of reviewer 1

Congratulations to the authors for such a good work. I would like to share some comments/seek for some clarification.

1. Suggest authors to expand on the importance and the use of quality indicators (QIs) in health services. It is unclear in the introduction why authors want to identify if the PROMs can discriminate PT practices.

Author’s reply:

Thank you for your suggestion. We have added a paragraph providing a definition of QIs, showing their potential/importance regarding transparency of service delivery in healthcare. See the paraphrased text passage below.

“QIs are “measurement tools, screens, or flags that are used as guides to monitor, evaluate, and improve the quality of patient care, clinical support services, and organizational functions that affect patient outcomes”. 11 (p 524) QIs consist of explicitly defined and measurable items referring to the structures, processes or outcomes of care, 12,13,14 and have the potential to be used to support quality improvement, accountability and to provide transparency of service delivery in healthcare.15” [Line 57-62]

In addition we have added a sentence in the Introduction, explaining why we want to identify if the PROMs can be used to discriminate between PT practices. PROMs are used for the development of QIs since they enable the measurement of health outcomes based on the direct perception of patients and these instruments are recommended in the Clinical Practice Guidelines (CPGs) of KNGF.

“PROMs were chosen for the development of QIs, since they enable the measurement of health outcomes based on the direct perception of patients and these instruments are recommended in the high quality CPGs of KNGF.” [Line 79-81]

2. Table III - Suggest to have the total N in the top row and present the data in the format of N(%). Suggest to use age categories of similar gap, if not, please justify in Materials and methods section.

Author’s reply:

Thank you for your suggestion. We have changed the order of presentation in Table III. For both outcomes (‘Age distribution’ and ‘Gender distribution’) we now first present the total N (PT-practices and patients). In addition we present the data in the suggested format: N (%). [Line 293-297]

Unfortunately we are not able to change the presented age categories because we have used the age categories as used in the selected national reference data. Since we cannot change the age categorisation in this national reference data, we are obliged to use the categorisation as they are presented in their current form. We have provided the following explanation in the Materials and methods section

“The age categories chosen for this study were aligned with those used in the national reference dataset.” [Line 180-181]

3. Table IV, V, VI - how does the difference being derived? For example, mean age difference has a difference in %. Suggest author to include the explanation in "Materials and methods" section.

Author’s reply:

Thank you for your suggestion. In the “Materials and methods” section we have added the sentence as paraphrased below, explaining how we have derived the differences as presented in the Tables IV, V and VI. Since similar differences have been presented in Table III we have also included this explanation for this Table.

“Differences between the created categories were established by calculating the relative differences on the outcomes of the created groups.” [Line 203-205 & 218-220]

In addition we have added a footnote to the Tables, explaining how the differences were derived. [Line 295, 313-314, 319-320, 333]

4. Kindly specify test used to obtain p-values for Table III - VI. Is it t-test or Fisher's exact test?

Author’s reply:

Thanks you for your suggestion. In the text and in the footnotes of the Tables we have specified that we used a t-test to calculate the p-values. [Line 199, 216, 296, 321]

5. Please provide the number of PT practices that were included in the analysis. Authors only described the number of patients .

Author’s reply:

Thank you for your final comment. We have provided the number of PT practices in Table I [Line 273-277], III [Line 293-297], the description of Table IV [Line 308-311] V [ Line 316-322] & VI [Line 331-337]. In addition throughout the text we have described more clearly what data selection was made for each analysis, providing insight why the numbers (of PT practices and patients) differ in the different Tables [Line: 163, 176, 180-181, 225, 265-266, 282, 284-285, 287, 293, 299]

Comments of reviewer 2

Comments to the authors

Manuscript title:

Reliability, validity and discriminability of patient reported outcomes for non-specific low back pain in a nationwide physical therapy registry: a retrospective observational cohort study

Thank you for inviting me to review the manuscript which aimed at testing to what extent health outcomes in patients with non-specific low back pain (NSLBP) collected with PROMs in the national clinical registry of the Royal Dutch Society for Physical Therapy (KNGF) are reliable, valid and able to discriminate between physical therapy practices.

The paper provides background information about successful involvement of patients and physical therapists in a clinical registry in The Netherlands. The authors included data from an impressive number of physical therapy (PT) practices (n=865) and PTs (n=6,560) of patients with NSLBP over a long period of time, between 2013 and 2018. The comprehensive description of included outcome measures in addition to the methods used to answer the research questions were helpful. Findings were described in detail and discussed sufficiently.

This manuscript addresses an important issue for clinicians, researchers and other stakeholders as it demonstrates the importance and value of clinical registries also for formulating appropriate QI's and provides a good example of a practice test to assess the psychometric properties of the collected data. Therefore, it should be of interest to a wide audience.

I have some minor suggestions that could improve the manuscript further.

1. Abstract: The abstract summarizes the study very well. Please just correct “in this next stages of implementation”.

Author’s reply:

Thank you for your correction with regard to this grammatical error. We have changed wording and used “the” instead of “this”. [Line 44]

2. Introduction: The introduction is comprehensive and informs the reader about the importance of clinical registries also for the development of quality indicators. It would be interested how people consented to participate in the registry and if they have to consent only once or multiple times.

Author’s reply:

Thank you for your suggestion to include a description on how the informed consent is obtained from both the patient and the therapist in the Introduction. We have added the following sentence in the Introduction:

“Data are only collected after a one-off informed consent was provided by the patient and their therapist and recorded in the patient file.” [Line 69-70]

Besides adding this sentence in the Introduction we felt our explanation on how the informed consent was obtained, as described in the Materials and methods section, needed some clarification. Therefore we have reworded the initial sentence in the Materials and methods section as paraphrased below:

“Data were only uploaded to the national clinical registry from the EHR-systems if: 1) the patient provided their informed consent for anonymous use of the clinical data from their patient record, and 2) the therapist provided their informed consent on the use of all patient records for which they received an informed consent of the patient.” [Line 111-114]

3. Method: described comprehensively. In this manuscript, acute vs. chronic NSLBP is defined with acute being 0-12 weeks since onset, and chronic >12 weeks. However, definitions of acute and chronic LBP are different, for example in a review provided by Chou et al. where acute NSLBP is defined <4 weeks' duration) while chronic/subacute NSLBP >4 weeks' duration. Please provide a reason for your definition and a reference.

Chou, R., & Huffman, L. H. (2007). Nonpharmacologic therapies for acute and chronic low back pain: a review of the evidence for an American Pain Society/American College of Physicians clinical practice guideline. Annals of internal medicine, 147(7), 492-504.

Author’s reply:

We understand this question. Indeed, in the literature there are different definitions for patients with (sub)acute versus those with chronic NSLBP. We choose to use the definition provided by the paper of Costa et al. since their meta-analysis from 2012 confirmed the hypothesis that there is a difference in the recovery of patients with (sub)acute versus those with chronic complaints. We have added the reference to this paper in the ‘Data collection’ paragraph in the Materials and methods section.

“…the level of chronicity of their complaints was recorded (i.e. (sub)acute vs. chronic, with (sub)acute being 0-12 weeks since onset, and chronic >12 weeks29)…” [Line 118-120]

29. Costa L da CM, Maher CG, Hancock MJ, McAuley JH, Herbert RD, Costa LOP. The prognosis of acute and persistent low-back pain: a meta-analysis. Can Med Assoc J. 2012;184(11):1229-1230. doi:10.1503/cmaj.120627

In addition we decided to be more precise in declaring the included subgroups. In accordance with the paper of Costa et al. we have changed the wording ‘acute’ to ‘(sub)acute’ throughout the entire article.

4. Results: are very well described. The tables provided help to overlook the results, but please look at the formatting of the tables, e.g. upper and lower lines in Table 1 missing. Table III By “total clinical registry” you mean all data of the KNGF’s clinical registry?

Author’s reply:

Thank you for the provided feedback. We have checked the formatting of the Tables and changed the wording of Table III as suggested. [Line 293-297]

5. Discussion: very well written. I liked the idea of “implementation activities” to facilitate the use of the registry.

Line 295 “that the completions of a repeated PROM measurement has no added value for this patient category” should be have…

Author’s reply:

Thank you for your correction with regard to this grammatical error. We have changed wording and used “have” instead of “has”. [Line 368]

6. Further suggestions: Look for double spaces and commas throughout the text (e.g. after “In addition”, “Nevertheless”).

Author’s reply:

Thank you for your final correction. We have deleted all double spaces and added the commas when necessary.

Decision Letter 1

Alison Rushton

5 May 2021

Reliability, validity and discriminability of patient reported outcomes for non-specific low back pain in a nationwide physical therapy registry: a retrospective observational cohort study

PONE-D-21-03296R1

Dear Dr. Meerhoff,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Alison Rushton

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Thank you for addressing all of the feedback from the two reviewers to a satisfactory level.

Reviewers' comments:

Acceptance letter

Alison Rushton

25 May 2021

PONE-D-21-03296R1

Reliability, validity and discriminability of patient reported outcomes for non-specific low back pain in a nationwide physical therapy registry: a retrospective observational cohort study

Dear Dr. Meerhoff:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Professor Alison Rushton

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix. The calculation used to define a more specific sample size, using the equation of Twisk et al. (2013).

    (DOCX)

    S1 Fig. A flowchart of the patient inclusion.

    (DOCX)

    S1 Syntax. 20180309_cleaning_definitief_GUUS_JC.sps.

    (SPS)

    S2 Syntax. 20201603 descriptives and provisionary analysis and QI analysis.sps.

    (SPS)

    S1 Dataset. 20180116_Totaalbestand_aanlever.sav.

    (ZIP)

    S2 Dataset. werkbestand_20180308_DCSPH.sav.

    (ZIP)

    S3 Dataset. werkbestand_20180308_DCSPH_leeftijd.sav.

    (ZIP)

    S4 Dataset. werkbestand_20180308_DCSPH_leeftijd_geslacht.sav.

    (ZIP)

    S5 Dataset. werkbestand_20180308_DCSPH_leeftijd_geslacht_afgesloten.sav.

    (ZIP)

    S6 Dataset. werkbestand_20180308_DCSPH_leeftijd_geslacht_afgesloten_duurklachten.sav.

    (ZIP)

    S7 Dataset. 20180314_multilevel.sav.

    (ZIP)

    S8 Dataset. One sample T-test Nivel data vs LDF data.sav.

    (ZIP)

    Data Availability Statement

    We are happy to be able to share the anonymized dataset with PLOS ONE. This enables us to comply with the important academic FAIR principles. We have added all necessary datasets and syntaxes as Supporting Information files. In uploading the datafiles (SPSS) we had to compress (.ZIP) the files.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES