Skip to main content
The BMJ logoLink to The BMJ
. 2022 Jun 23;377:e063064. doi: 10.1136/bmj-2020-063064

The evidence base for US joint commission hospital accreditation standards: cross sectional study

Sarah A Ibrahim 1,2, Kelly A Reynolds 3, Emily Poon 2, Murad Alam 4,5,6,
PMCID: PMC9215261  PMID: 35738660

Abstract

Objective

To evaluate the evidence upon which standards for hospital accreditation by The Joint Commission on Accreditation of Healthcare Organizations (the Joint Commission) are based.

Design

Cross sectional study.

Setting

United States.

Participants

Four Joint Commission R3 (requirement, rationale, and reference) reports released by July 2018 and intended to become effective between 1 July 2018 and 1 July 2019.

Interventions

From each R3 report the associated standard and its specific elements of performance (or actionable standards) were extracted. If an actionable standard enumerated multiple requirements, these were separated into distinct components. Two investigators reviewed full text references, and each actionable standard was classified as either completely supported, partly supported, or not supported; Oxford evidence quality ratings were assigned; and the Grading of Recommendations, Assessment, Development, and Evaluation (GRADE) was used to assess the strength of recommendations.

Main outcome measure

Strengths of recommendation for actionable standards.

Results

20 actionable standards with 76 distinct components were accompanied by 48 references. Of the 20 actionable standards, six (30%) were completely supported by cited references, six were partly supported (30%), and eight (40%) were not supported. Of the six directly supported actionable standards, one (17%) cited at least one reference of level 1 or 2 evidence, none cited at least one reference of level 3 evidence, and five (83%) cited references of level 4 or 5 evidence. Of the completely supported actionable standards, strength of recommendation in five was deemed GRADE D and in one was GRADE B.

Conclusions

In general, recent actionable standards issued by The Joint Commission are seldom supported by high quality data referenced within the issuing documents. The Joint Commission might consider being more transparent about the quality of evidence and underlying rationale supporting each of its recommendations, including clarifying when and why in certain instances it determines that lower level evidence is sufficient.

Introduction

The Joint Commission on Accreditation of Healthcare Organizations is a national body that accredits and evaluates more than 22 000 facilities in the United States, including hospitals, behavioral health centers, nursing homes, home health services, ambulatory care centers, and other healthcare organizations across the United States.1 2 Its mission is to promote patient safety and improve quality of healthcare organizations by enforcing compliance with various performance based standards.3

Although accreditation by The Joint Commission is voluntary, meeting mandated standards deems healthcare organizations eligible to receive federal reimbursement from the Centers for Medicare and Medicaid Services.4 Many states also include The Joint Commission standards in their hospital licensure requirements.2 5 Failure to meet the standards therefore could potentially disqualify a hospital from receiving millions of dollars annually in federal funding and could even lead to suspension or revocation of its state license.2

In recent years, advisory and regulatory bodies have increasingly accepted that the evidence underlying recommendations in clinical practice guidelines should be conveyed. In 2011 the US Institute of Medicine published a guidance document for the development of trustworthy, evidence based guidelines in which it recommended that organizations producing guidelines provide clear, transparent explanations of the reasoning supporting their published standards.6 Specifically, the Institute of Medicine recommended the inclusion of a discussion of expected benefits and harms; a summary of the available evidence, including its quality, completeness, and consistency, as well as any relevant evidence gaps; and details about the influence of other types of information, such as values, opinion, theory, and clinical experience, that might have influenced the recommendations. This would culminate in a rating of the level of confidence in the supporting evidence and also a rating of the strength of recommendation associated with guidelines or standards. Since the publication of the Institute of Medicine guidance, the US Preventive Services Taskforce has published standards for guideline development that accepts the Institute of Medicine’s guidance as the basis for establishing evidence foundations for its own recommendations.7

Clinical practice guidelines are not standards. Both the Institute of Medicine and the US Preventive Services Taskforce focus on clinical practice guidelines. A guideline recommendation is a suggestion based on evidence. An actionable standard, such as that implemented by The Joint Commission, is a requirement, not a suggestion. Arguably, the level of evidence and transparency should be greater moving along this continuum, from clinical practice guidelines to standards. Given that The Joint Commission accreditation is a deemed authority for federal certification and is therefore a de facto national standard, it is reasonable to expect that the evidentiary basis underlying the development of The Joint Commission’s standards meet and exceed the evidentiary basis considered appropriate for mere clinical practice guidelines.8 Between July 2018 and June 2020 we evaluated the quality of the evidence on which The Joint Commission standards are based and the extent to which discussion of this information is transparent and available.

Methods

Definitions of The Joint Commission’s terminology

Standards—The Joint Commission standards guide the evaluation of safety and assurance of quality of care and are the basis of its accreditation process. Standards according to The Joint Commission are broad thematic areas requiring oversight. Examples include pain assessment and pain management, identifiers for newborns, and assessment of infection in the perinatal period.9 10 11 12 13 Developed in response to emerging safety and quality concerns, standards and their subsections are drafted with stakeholder input and then distributed nationally for field review.9 After review, executive leadership of The Joint Commission might revise the standards before publication and implementation.9 Manuals describing all standards are distributed to accredited institutions and can also be purchased.2 9 14 The Joint Commission currently maintains more than 250 standards, and it regularly updates these with variable frequency.3 15

Elements of performance (actionable standards)—Element of performance is The Joint Commission term for an actionable standard on which basis healthcare organizations are evaluated. Although the term standard is used to denote general thematic areas for compliance, specific actionable standards (elements of performance) within these broad categories are the activities assessed by site reviewers and are used to score compliance with regulations.14

R3 reports—R3 refers to requirement, rationale, and reference. R3 reports are The Joint Commission’s documents that announce and describe new or revised standards, provide the background underlying these standards, and list literature citations on which each standard’s specific elements of performance (actionable standards) are based. R3 reports are distributed to The Joint Commission accredited healthcare organizations and are also made publicly available on its website.16

Selection criteria

We included all R3 reports released by July 2018 and intended to become effective between 1 July 2018 and 1 July 2019 and excluded standards outside this timeframe, as well as duplicate standards. Actionable standards that cited the same literature and therefore had similar levels of support were also excluded. Within each group of similar actionable standards, we used the actionable standard that had the highest degree of support and strongest level of evidence for this analysis.

Data extraction

From each R3 report we extracted each associated standard as well as its specific actionable standards. If an actionable standard seemed to enumerate multiple requirements (eg, need for multiple diagnostic tests, roles of different types of providers, or for different classes of patients), two reviewers (SI, KR) used qualitative research methods to separate these actionable standards into distinct components (supplementary table 1). Specifically, any actionable standard that described multiple actions or multiple agents responsible for adhering to the actionable standard was divided so that a combination of only one action and one agent were included in each requirement.

Support provided by cited references

Two investigators (SI, KR) independently reviewed the references cited in each actionable standard and classified the degree to which each reference supported the corresponding distinct component as one of: completely supported (the reference literally or substantially affirmed the distinct component); partly supported (the reference addressed the same general topic as the distinct component but did not literally or substantially affirm the distinct component—for example, if a distinct component required testing of mothers for HIV and the reference was related to HIV in pregnant women but did not mention testing, the reference was classified as partly supports); or no support (the reference contained no language, explicit or implied, to support the distinct component). The category for partly supported actionable standard was subjective but considered necessary so as to avoid minimizing the utility of relevant references by excluding those at least somewhat relevant. The investigators were blinded to each other’s preliminary ratings. To improve reliability, the investigators completed their ratings using a standardized checklist, adapted from the Centre for Evidence-Based Medicine Critical Appraisal Tools (supplementary table 2). A third rater (MA) resolved discrepancies, with the highest level of support (ie, rounding up) reported by at least two raters selected.

If the distinct components within an actionable standard were totally supported by the references, then the actionable standard was considered completely supported. If at least one distinct component was supported by the references, then the actionable standard was considered partly supported. If none of the distinct components were supported by the references, then the actionable standard was considered not supported.

Level of evidence for cited references

We assigned a level of evidence to each reference using the Oxford Centre for Evidence-Based Medicine scheme.17 This tool is commonly used to appraise evidence based reviews and clinical practice guidelines and was chosen because, unlike other rating schemes, it permits a higher level of evidence for some observational studies.18

We modified the tool using two methods. First, if a reference was a clinical practice guideline, we evaluated the recommendation’s level of evidence within the guideline rather than the guideline as a whole. We did this because although guidelines are collectively considered expert opinion, some guideline recommendations might be supported by high level evidence and others might be based solely on lower level evidence, such as expert opinion. Second, we assigned cross sectional studies—which are not well specified in the Oxford tool—an evidence level of 4.

Two investigators (SI, KR) independently assigned a level of evidence to each reference. A third investigator (MA) resolved discrepancies.

Assessment of overall strength of recommendations

We used a modified GRADE (Grading of Recommendations, Assessment, Development, and Evaluation) approach to assess the strength of the actionable standards, accounting for the quality of each supporting reference.19 Two investigators (SI, KR) independently assigned a GRADE rating to each actionable standard; discrepancies were resolved by using the evaluation of a third investigator (MA).

Statistical analysis

Data were extracted and saved in Microsoft Excel. Analyses were carried out in Excel and verified in R (version 4.1.0). Descriptive statistics were calculated on the actionable standards, and their distinct components reviewed, including counts and percentages. We used unweighted Cohen’s κ statistics to evaluate interrater reliability of reviewers’ assessments of the level of evidence and degree of support.20

Patient and public involvement

Speaking to patients inspired this review. Although no patients or members of the public were directly involved in this paper owing to the methodological focus on assessing quality of evidence, we did speak to patients about the study, and we asked a member of the public to read our manuscript.

Results

Four of the five R3 reports issued by July 2018 were included (fig 1). One report was excluded as it was a duplicate for implementation in a different clinical care setting. One report each was developed for use in ambulatory care settings, obstetrics services, newborn patient services, and critical access hospitals. For a total of 23 actionable standards, each report contained between one and 19 (median 7) actionable standards. The actionable standards element of performance 14 and element of performance 15 were excluded because they addressed similar clinical care decisions and contained similar reference lists to element of performance 16. In addition, element of performance 6 (identifier PC.01.02.07) was excluded owing to its similarity to element of performance 7 (LD.04.03.13). As a result, 20 actionable standards were included in the final analysis.

Fig 1.

Fig 1

Flowchart showing total number of The Joint Commission on Accreditation of Healthcare Organizations R3 (requirement, rationale, and reference) reports, actionable standards, and distinct components of actionable standards analyzed, and number of distinct components that were completely, partly, or not supported by the evidence

We separated the 20 actionable standards into 76 distinct components. Across the 76 components, 48 references were cited. One distinct component was excluded from subsequent analyses because no references were cited. One reference was a poster presentation and was not publicly accessible and was therefore excluded from this analysis.

Each distinct component analyzed cited between one and 10 references (median 2, interquartile range 1.0-2.5). Of the 75 distinct components, 68% (n=51) were supported by at least one reference, 31% (n=23) were partly supported by at least one reference, and 17% (n=13) were not supported by references. The interrater reliability for the two raters was substantial (κ=0.77, 95% confidence interval 0.70 to 0.83; P<0.001). Of the 244 assessments for degree of support of a reference, the primary reviewers disagreed in 36 instances (15%). Thirty (83%) of these adjudications resulted in rounding up.

Of the 47 references evaluated, most were low quality and were classified as evidence level 4 or 5 (n=34, 72%) (table 1). Among the distinct components directly supported by one or more references, 51% (26/51) were supported exclusively by references with evidence quality ratings of 5. Seven of the 19 actionable standards (37%) that cited references were directly supported only by narrative reviews, opinion based references, or book excerpts. Fewer than 30% of references were of evidence quality 1 or 2 (12/47 references), and only 13 of 51 distinct components (25%) were directly supported by such evidence (ie, systematic reviews of randomized controlled trials or other high quality studies). The interrater reliability for the two raters was substantial (κ=0.73, 95% confidence interval 0.66 to 0.81; P<0.001). Of the 51 unique assessments for level of evidence of a reference as it pertained to a distinct component, the primary reviewers disagreed in 11 instances (22%). Six (55%) of these adjudications resulted in rounding up.

Table 1.

Level of evidence associated with actionable standards

Actionable standards References in R3 report Quality of evidence*
PC.01.02.01, New EP 16: For critical access hospitals that provide obstetric services: if the mother tests positive for HIV, hepatitis B, group B streptococcus, or syphilis when tested in labor and during delivery, that information is also documented in the newborn’s medical record after delivery 40, 41, 42, 43, 44, 45, 46, 47, 48, 49 5, 5, 5, 5, 5, 4, 2c, 5, 5, 2b
LD.04.03.13, EP 1: The organization has a leader or leadership team that is responsible for pain management and opioid prescribing and developing and monitoring performance improvement activities 50, 51 5, 2b
LD.04.03.13, EP 3: The organization provides staff and licensed independent practitioners with educational resources to improve pain assessment, pain management, and the safe use of opioid drugs based on the identified needs of its patient population 52, 53 1a, 2b
LD.04.03.13, EP 4: The organization provides information to staff and licensed independent practitioners on available services for consultation and referral of patients with complex pain management needs 54, 55 5, 5
LD.04.03.13, EP 6: The organization facilitates practitioner and pharmacist access to the Prescription Drug Monitoring Program databases 56, 57 5, 5
PC.01.02.07, EP 1: The organization has defined criteria to screen, assess, and reassess pain that are consistent with the patient’s age, condition, and ability to understand 58, 59, 60, 61, 62 5, 5, 5, 5, 5
PC.01.02.07, EP 3: The organization treats the patient’s pain or refers the patient for treatment 54, 63, 64 3a, 4, 5
PC.01.02.07, EP 4: The organization develops a pain treatment plan based on evidence based practices and the patient’s clinical condition, medical history, and pain management goals 54, 60, 65, 66 3a, 5, 5, 5
PC.01.02.07, EP 5: The organization involves patients in the pain management treatment planning process through the following: developing realistic expectations and measurable goals that are understood by the patient for the degree, duration, and reduction of pain; discussing the objectives used to evaluate treatment progress; providing education on pain management, treatment options, and safe use of opioid and non-opioid drugs when prescribed 67 4
PC.01.02.07, EP 7: Based on the patient’s condition, the organization reassesses and responds to the patient’s pain through: evaluation and documentation of responses to pain interventions; progress toward pain management goals, including functional ability; side effects of treatment; risk factors for adverse events caused by the treatment 60, 62 5, 5
PC.01.02.07, EP 8: The organization educates the patient and family on discharge plans related to pain management including the following: pain management plan of care; side effects of pain management treatment; if applicable, activities of daily living, including the home environment, that might exacerbate pain or reduce effectiveness of the pain management plan of care, as well as strategies to address these issues; safe use, storage, and disposal of opioids when prescribed 68 4
PI.02.01.01, EP 19: The organization monitors the use of opioids to determine if they are being prescribed safely 69, 70 5, 4
NPSG.01.01.01, EP 3: For newborn patients: use distinct methods of identification for newborn patients 71, 72, 73, 74, 75, 76 2b, NA, 5, 4, 5, 4
LD.04.03.13, EP 2: The critical access hospital provides non-drug pain treatment modalities 54, 77, 78, 79, 80, 81 1a, 2b, 1a, 5, 5, 2b
LD.04.03.13, EP 7: Critical access hospital leadership works with its clinical staff to identify and acquire the equipment needed to monitor patients who are at high risk for adverse outcomes from opioid treatment 82, 83, 84 3a, 5, 2a
MS.05.01.01, EP 18: The medical staff is actively involved in pain assessment, pain management, and safe opioid prescribing through the following: participating in the establishment of protocols and quality metrics; reviewing performance improvement data 50 5
PC.01.02.07, EP 2: The critical access hospital screens patients for pain during emergency department visits and at the time of admission 62 5
LD.04.03.13, EP 5: The critical access hospital identifies opioid treatment programs that can be used for patient referrals No references were listed  
PI.01.01.01, EP 40: The critical access hospital collects data on pain assessment and pain management including types of interventions and effectiveness 85 5
PI.02.01.01, EP 18: The critical access hospital analyzes data collected on pain assessment and pain management to identify areas that need change to increase safety and quality for patients 86, 87 5, 2c

R3=requirement, rationale, and reference; EP=element of performance.

Evidence level of 1 indicates high level of evidence, and 5 indicates lowest level of evidence—specifically, individual randomized controlled trial with narrow confidence interval, systematic review with homogeneity of randomized controlled trials; well designed controlled trial without randomization, prospective comparative cohort trial, systematic review with homogeneity of cohort studies; systematic review with homogeneity of level 3 and better studies, case-control study, retrospective cohort study; case series with or without intervention, cross sectional study, poor quality cohort and case-control studies; expert opinion without explicit critical appraisal, case reports.19

*

If a reference was a clinical practice guideline, level of evidence was evaluated for the recommendation within the guideline, rather than the guideline as a whole.

Of 20 actionable standards, six (30%) were completely supported by references, six were partly supported (30%), and eight (40%) were not supported (table 2). Of the six completely supported actionable standards, one (17%) cited at least one reference of level 1 or 2 evidence, none cited at least one reference of level 3 evidence, and five (83%) cited references of level 4 or 5 evidence. Strength of recommendation of the completely supported actionable standards was GRADE B in one and GRADE D in five (table 3). Of the six actionable standards only partly supported, two (33%) cited at least one reference of level 1 or 2 evidence, one (17%) cited level 3 evidence, and three (50%) exclusively cited references of level 4 or 5. Strength of recommendation of the partly supported actionable standards was GRADE C in three (50%) and GRADE D in three (50%).

Table 2.

Extent of direct support from references for actionable standards

Actionable standards No of distinct components Extent of support from references
PI.02.01.01, EP 19: The organization monitors the use of opioids to determine if they are being prescribed safely 1 Directly supported
PI.01.01.01, EP 40: The critical access hospital collects data on pain assessment and pain management, including types of interventions and effectiveness 2 Directly supported
PC.01.02.07, EP 7: Based on the patient’s condition, the organization reassesses and responds to the patient’s pain through the following: evaluation and documentation of responses to pain interventions; progress toward pain management goals, including functional ability 7 Directly supported
PC.01.02.07, EP 2: The critical access hospital screens patients for pain during emergency department visits and at the time of admission 2 Directly supported
LD.04.03.13, EP 6: The organization facilitates practitioner and pharmacist access to the Prescription Drug Monitoring Program databases 2 Directly supported
LD.04.03.13, EP 2: The critical access hospital provides non-drug pain treatment modalities 1 Directly supported
PC.01.02.07, EP 5: The organization involves patients in the pain management treatment planning process through the following: developing realistic expectations and measurable goals that are understood by the patient for the degree, duration, and reduction of pain; discussing the objectives used to evaluate treatment progress; providing education on pain management, treatment options, and safe use of opioid and non-opioid drugs when prescribed 10 Partly supported
PC.01.02.07, EP 4: The organization develops a pain treatment plan based on evidence based practices and the patient’s clinical condition, medical history, and pain management goals 4 Partly supported
MS.05.01.01, EP 18: The medical staff is actively involved in pain assessment, pain management, and safe opioid prescribing through the following: participating in the establishment of protocols and quality metrics; reviewing performance improvement data 6 Partly supported
LD.04.03.13, EP 4: The organization provides information to staff and licensed independent practitioners on available services for consultation and referral of patients with complex pain management needs 4 Partly supported
LD.04.03.13, EP 3: The organization provides staff and licensed independent practitioners with educational resources to improve pain assessment, pain management, and the safe use of opioid drugs based on the identified needs of its patient population 5 Partly supported
LD.04.03.13, EP 1: The organization has a leader or leadership team that is responsible for pain management and opioid prescribing and developing and monitoring performance improvement activities 4 Partly supported
PC.01.02.07, EP 8: The organization educates the patient and family on discharge plans related to pain management, including the following: pain management plan of care; side effects of pain management treatment; if applicable, activities of daily living, including the home environment, that might exacerbate pain or reduce effectiveness of the pain management plan of care, as well as strategies to address these issues; safe use, storage, and disposal of opioids when prescribed 11 Not supported
PC.01.02.07, EP 3: The organization treats the patient’s pain or refers the patient for treatment 2 Not supported
PC.01.02.07, EP 1: The organization has defined criteria to screen, assess, and reassess pain that are consistent with the patient’s age, condition, and ability to understand 5 Not supported
NPSG.01.01.01, EP 3: For newborn patients: Use distinct methods of identification for newborn patients 1 Not supported
PC.01.02.01, New EP 16: For critical access hospitals that provide obstetric services: if the mother tests positive for HIV, hepatitis B, group B streptococcus, or syphilis when tested in labor and during delivery, that information is also documented in the newborn’s medical record after delivery 4 Not supported
LD.04.03.13, EP 7: Critical access hospital leadership works with its clinical staff to identify and acquire the equipment needed to monitor patients who are at high risk for adverse outcomes from opioid treatment 2 Not supported
LD.04.03.13, EP 5: The critical access hospital identifies opioid treatment programs that can be used for patient referrals 1 Not supported
PI.02.01.01, EP 18: The critical access hospital analyzes data collected on pain assessment and pain management to identify areas that need change to increase safety and quality for patients 2 Not supported

EP=element of performance.

Table 3.

Strength of recommendation for actionable standards

Actionable standards GRADE strength of recommendation
LD.04.03.13, EP 2: The critical access hospital provides non-drug pain treatment modalities B
PC.01.02.07, EP 4: The organization develops a pain treatment plan based on evidence based practices and the patient’s clinical condition, medical history, and pain management goals C
PC.01.02.07, EP 3: The organization treats the patient’s pain or refers the patient for treatment C
NPSG.01.01.01, EP 3: For newborn patients: use distinct methods of identification for newborn patients C
LD.04.03.13, EP 7: Critical access hospital leadership works with its clinical staff to identify and acquire the equipment needed to monitor patients who are at high risk for adverse outcomes from opioid treatment C
LD.04.03.13, EP 3: The organization provides staff and licensed independent practitioners with educational resources to improve pain assessment, pain management, and the safe use of opioid drugs based on the identified needs of its patient population C
LD.04.03.13 EP 1: The organization has a leader or leadership team that is responsible for pain management and opioid prescribing and developing and monitoring performance improvement activities C
PI.02.01.01, EP 19: The organization monitors the use of opioids to determine if they are being prescribed safely D
PI.01.01.01, EP 40: The critical access hospital collects data on pain assessment and pain management, including types of interventions and effectiveness D
PC.01.02.07, EP 8: The organization educates the patient and family on discharge plans related to pain management, including the following: pain management plan of care; side effects of pain management treatment; if applicable, activities of daily living, including the home environment, that might exacerbate pain or reduce effectiveness of the pain management plan of care, as well as strategies to address these issues; safe use, storage, and disposal of opioids when prescribed D
PC.01.02.07, EP 7: Based on the patient’s condition, the organization reassesses and responds to the patient’s pain through the following: evaluation and documentation of responses to pain interventions; progress toward pain management goals, including functional ability D
PC.01.02.07, EP 5: The organization involves patients in the pain management treatment planning process through the following: developing realistic expectations and measurable goals that are understood by the patient for the degree, duration, and reduction of pain; discussing the objectives used to evaluate treatment progress; providing education on pain management, treatment options, and safe use of opioid and non-opioid drugs when prescribed D
PC.01.02.07, EP 2: The critical access hospital screens patients for pain during emergency department visits and at the time of admission D
PC.01.02.07, EP 1: The organization has defined criteria to screen, assess, and reassess pain that are consistent with the patient’s age, condition, and ability to understand D
PC.01.02.01, New EP 16: For critical access hospitals that provide obstetric services: if the mother tests positive for HIV, hepatitis B, group B streptococcus, or syphilis when tested in labor and during delivery, that information is also documented in the newborn’s medical record after delivery D
MS.05.01.01, EP 18: The medical staff is actively involved in pain assessment, pain management, and safe opioid prescribing through the following: participating in the establishment of protocols and quality metrics; reviewing performance improvement data D
LD.04.03.13, EP 6: The organization facilitates practitioner and pharmacist access to the Prescription Drug Monitoring Program databases D
LD.04.03.13, EP 5: The critical access hospital identifies opioid treatment programs that can be used for patient referrals D
LD.04.03.13, EP 4: The organization provides information to staff and licensed independent practitioners on available services for consultation and referral of patients with complex pain management needs D
PI.02.01.01, EP 18: The critical access hospital analyzes data collected on pain assessment and pain management to identify areas that need change to increase safety and quality for patients D

GRADE=Grading of Recommendations, Assessment, Development, and Evaluation; EP=element of performance.

A indicates a strong recommendation with high quality evidence, such that further research is very unlikely to change confidence in the estimate of effect; B recommendation is based on moderate quality evidence, and further research is likely to have an important impact in confidence in the estimate of effect; C recommendation is based on low quality evidence, such that further research is very likely to change the estimate of effect; D recommendation is based on very low quality evidence, and any estimate of effect is very uncertain.21 The strength of these recommendations were generated using only the references provided in the R3 report.

Discussion

In general, when the Joint Commission on Accreditation of Healthcare Organizations implements new actionable standards for healthcare organizations, it often provides few references and little evidence in support of these standards in documents that are publicly available. When evidence is cited, it is often of low level or only partly supports the new actionable standards. As a consequence, it is unclear whether implementation of new actionable standards from The Joint Commission would likely improve safety or quality outcomes. Although The Joint Commission may have rationales that are not conveyed in its thinly referenced public documents, greater transparency about such additional reasonings as well as more complete reference lists may further strengthen confidence in the recommendations. Finally, if in some instances The Joint Commission believes evidence may not be required to support an actionable standard, then explicitly mentioning this, and explaining the rationale, could provide convincing justification in place of evidence.

Comparison with other studies

Although research on the detrimental effects of standards that are insufficiently supported by evidence is limited, there is a body of research on the possible harms from clinical practice guidelines that are inadequately supported. Despite standards being requirements and guidelines being recommendations, it is reasonable to expect standards to be at least as well supported as guidelines.

A body of research confirms the potentially deleterious effects of unsubstantiated, low quality recommendations in clinical practice guidelines on patient care. A review of articles published in a high impact journal found that over a span of 10 years, medical practice was reversed in 146 cases after the publication of one or more higher quality trials refuted the intended goal of a recommended practice, or found that benefits were outweighed by potential harms.22 However, once practices have been implemented, albeit with lower level evidence, reversal can be difficult.21 To avoid implementation of unwarranted practices and delay in their reversal, many clinicians and researchers believe that new guidelines should be adopted only after the availability of rigorous evidence supporting effectiveness, affordability, and practicality.20 21 22 23 24

Even if enforcement of recommendations unsupported by evidence does not directly harm patients, it might lead to regulatory fatigue that diverts clinicians and facilities from their primary missions.21 25 26 27 28 29 Efforts to comply could distract workers from implementing policies that are essential, and possibly result in patient harm.25 Time and resources could also be wasted. Fewer recommendations founded on higher quality evidence may help alleviate regulatory fatigue and improve patient outcomes.25 30

This study detected several types of evidentiary problems. In some cases, The Joint Commission may have created overly broad standards that go beyond the evidence, with this possibly based on internal considerations that are not made clear. In other instances, the supporting literature appears to be tangentially or minimally relevant. This latter problem may be attributable to lack of attention or the unavailability of relevant evidence—although at least in some cases relevant citations might have been inadvertently overlooked. For example, during the analysis of this study, the authors came across published high level evidence that was not cited in support of particular requirements.

Not only can outside observers be confused about the steps involved in The Joint Commission’s development process for standards, but in some cases the lack of transparency of this process might also be difficult for The Joint Commission to manage internally, thus resulting in directives that may not be aligned with patients’ best interests. For illustration, in 2002 The Joint Commission published measures requiring clinicians to obtain blood cultures before administering antibiotics to patients with community acquired pneumonia. It was subsequently found that the standards directly opposed the evidence already established in the literature, and transparency about the evidence that had informed the recommendations was minimal.31 Similarly, in 2000 The Joint Commission produced standards to address the problem of undertreatment of pain. These standards were founded on small, low level evidence studies that suggested a benefit from adhering to these standards, and subsequent reports of copious adverse events related to overtreatment of pain soon began to emerge.32

It is possible that literature supporting some recommendations by The Joint Commission cannot be found, given the nature and type of recommendation. In such cases, it would be prudent either to qualify the recommendation as a suggestion rather than a requirement or to narrow down the recommendation so that it can be supported by evidence. Greater specificity in recommendations and their evidentiary basis may also motivate increased compliance and make this more achievable.

Strengths and limitations of this study

This study has several limitations. We only assessed The Joint Commission R3 reports and only reviewed the references cited in each report. It is possible that The Joint Commission did not cite higher quality evidence that may have been available, including unpublished internal data. Indeed, each report includes the disclaimer, “not a complete literature review” so it is possible that the reports are not heavily referenced, and that additional supporting documentation or expert testimony is not disclosed. However, it seems unlikely that more appropriate published references were reviewed by The Joint Commission but were not included in the R3 reports, as abbreviated or abridged reference lists tend to include the strongest and most relevant references. In support of this, each R3 report includes the statement, “The references provide evidence that supports the requirement.” To the extent that The Joint Commission enforces de facto national standards, transparency in communicating the basis for new regulations is a reasonable expectation. So, if more convincing evidence supportive of The Joint Commission recommendations was available to the commission but not included in its R3 reports, The Joint Commission might consider modifying its practice in the future to share this additional evidence more fully with stakeholders. Further, if more relevant evidence exists, then by promptly and clearly sharing this evidence, The Joint Commission may also garner more stakeholder support for recommendations and spur more rapid adoption of recommendations.

Although we focused on assessing evidence in this study, we concede that evidence might not be required in all recommendations from The Joint Commission. In several conditions, evidence may be unnecessary or unobtainable—for example, when standards are based on self-evident or widely accepted imperatives for which acquisition of evidence is exceedingly difficult. Similarly, standards developed in response to unanticipated and emerging crises (eg, pandemics) may be implemented before evidence of their utility is determined, owing to the urgency of the situation. Standards may also be designed to prevent rare but extremely grave adverse outcomes, and because of the rarity of these adverse outcomes, it might be difficult to acquire data on the specific utility of the measures. Finally, standards developed to satisfy particular legal or regulatory mandates might not require evidence, as these standards are founded on governmental oversight authority. While there are thus several scenarios in which evidence may not be required for standards, it would, in each such case, be useful for agencies that set these standards to transparently admit the lack of supportive evidence and to explain on what basis these standards were being implemented, given the absence of evidence. Since even the most seemingly reasonable recommendations can have unintended negative consequences or take time and effort away from more useful activities, a strong argument for explaining the need for each standard is imperative.31 Indeed, The Joint Commission does provide a written rationale for each of its actionable standards, but at present these explanations do not routinely weigh the relative impact of evidence versus other bases for implementing standards.

Other limitations of this study include the choice to focus only on The Joint Commission standards. Standards can be used to ensure quality and safety of care but are not always appropriate for achieving healthcare safety goals. Sometimes, less directive standards—or even guidelines—might not be the answer. Specifically, quality improvement and improved patient safety could also require focused attention on adaptation and resilience in the context of uncertainty. Although standards work well to guard against known system vulnerabilities that lead to error, many uncertain system vulnerabilities and events might also lead to error. Unwanted events as a result of uncertainty are more difficult to mitigate because the causal mechanisms involved are unknown, and this represents a commonly unaddressed system safety issue in healthcare.33 In other words, when risk prevention strategies are well defined and objectively measurable, as in the case of ensuring proper patient identification, the implementation of particular standards may provide clear benefits. However, less quantifiable or predictable system issues, such as errors in interpersonal communication or cascading system failures (eg, cyberterrorism, simultaneous failure of different types of equipment, unanticipated acute supply shortages) may be addressed by strategies such as accident and incident analysis, restricting staff autonomy in high risk situations, encouraging monitoring and adaptation to the crisis, and in-the-moment problem solving by staff trained to expect and manage uncertainty.34 Healthcare workers must be trained and prepared accordingly.

Notably, this study focused only on The Joint Commission standards and a one year period during which newly implemented standards came into effect. The reason to avoid a historical assessment of standards was to emphasize current practice in The Joint Commission standard implementation, which has evolved over time. The cross sectional approach was also designed to pick up different types of standards, since by including all the R3 reports that met time criteria, bias associated with studying a particular content area, for which the level of evidence might not have been representative of The Joint Commission standards as a whole, was minimized. Nonetheless, our choices made could have led to a lack of representativeness about other nationwide or international quality and safety standards over the preceding decades.

Importantly, all but two of the actionable standards we reviewed considered acute or chronic pain. Thus, our findings might not be broadly applicable to all R3 reports by The Joint Commission. However, The Joint Commission should be expected to pay greater attention to the evidentiary basis behind pain management, given its experience in producing standards for the management of pain.

Policy implications

The direct and indirect costs to healthcare systems of adherence to accreditation rules are believed to be substantial. One systematic review of the cost of health accreditation, including accreditation by The Joint Commission, estimated that 0.2-1.7% of total annual healthcare system operating costs were spent on accreditation adherence.35 A case report estimated that it would cost an individual institution $326 784 (1% of the annual budget) to adhere to The Joint Commission specific accreditation standards.36 These statistics are likely underestimates because the indirect costs of adherence, which are likely to be much greater than the direct costs, are more difficult to measure and report.

Although the lack of transparency in the development process for standards is unsatisfactory for clinicians and researchers, it is not unusual, and it is mirrored in regulatory processes in other countries. For example, the International Society for Quality in Health Care External Evaluation Association (a separate entity established by the International Society for Quality in Health Care) provides a list of standards and criteria for adhering to these standards but does not provide supporting references.37 In contrast, the French National Authority for Health provides an accreditation handbook containing the standards it evaluates and notes that the development process for these standards includes a review of references pertaining to healthcare experiences as well as a consensus process. References were recently made available in the accreditation manual.38 Similarly, in the UK, the National Institute for Health and Care Excellence quality standards are accompanied by a topic library that collates evidence from NICE guidelines or NICE accredited evidence sources that was used by the Quality Standards Advisory Committee during the development of the quality standards.39

Conclusions

In general, recent actionable standards issued by The Joint Commission are not supported by high quality data referenced within the issuing documents. While performance standards play a critical role in ensuring high healthcare quality, and The Joint Commission remains an exemplary guardian of patients’ health and safety, The Joint Commission might consider being more transparent about the quality of evidence and underlying rationale supporting each of its recommendations. When higher level evidence is available to The Joint Commission but not cited in public documents, it might consider releasing this evidence publicly to increase confidence in its recommendations. In cases in which it believes lower level evidence (eg, unanimous expert consensus from a diverse group) could be sufficient to support a new actionable standard, or when it believes evidence might not even be necessary or appropriate (eg, a national emergency), a detailed, publicly available rationale would be helpful. Although these changes would increase work and cost, they might also increase enthusiasm for the uptake of The Joint Commission actionable standards, and further align the commission with the conventions of evidence based medicine. Finally, given its pivotal role in ensuring patient safety, The Joint Commission might want to increase its emphasis on the role of uncertainty in healthcare environments. Training staff to manage uncertainty through means other than standards and guidelines would complement the use of standards for more well defined and predictable problems.

What is already known on this topic

  • In the US, the Joint Commission on Accreditation of Healthcare Organizations mandates standards for safety in hospitals and other healthcare facilities

  • For each actionable standard, The Joint Commission provides a rationale, including supporting references

  • No study has systematically assessed the evidentiary basis underlying The Joint Commission’s actionable standards

What this study adds

  • Based on documents published by The Joint Commission in 2018 and 2019, six of 20 (30%) actionable standards were directly supported by the references provided, with most being assigned strength of recommendation GRADE D

  • The remaining actionable standards were either only partly supported (6/20, 30%), or not supported (8/20, 40%) suggesting that actionable standards recently issued by The Joint Commission are seldom supported by high quality data referenced in the issuing documents

  • The Joint Commission might consider being more transparent about its assessment of evidence, including providing more complete and convincing evidence for new standards, or explaining why lower level evidence might be sufficient in particular instances

Acknowledgments

We thank Bianca Y Kang for her invaluable contributions in extracting the data and in creating the manuscript figure and table. We also thank Jacob M Schauer for providing his expertise in statistical analysis.

Web extra.

Extra material supplied by authors

Supplementary information: Supplementary tables 1 and 2

ibrs063064.ww.pdf (119.1KB, pdf)

Contributors: All authors conceived and designed the study and acquired, analyzed, and interpreted the data. SAI, KAR, and MA drafted the manuscript. All authors critically revised the manuscript for important intellectual content. SAI and KAR performed the statistical analysis. MA obtained funding. SAI, KAR, and EP provided administrative, technical, or material support. MA supervised the study. MA had full access to the data in the study and takes responsibility for the integrity of the data and the accuracy of the data analysis. MA is the guarantor. The corresponding author attests that all listed authors meet authorship criteria and that no others meeting the criteria have been omitted.

Funding: This study was supported by the Department of Dermatology, Northwestern University. The funders had no role in considering the study design or in the collection, analysis, interpretation of data, writing of the report, or decision to submit the article for publication.

Competing interests: All authors have completed the ICMJE uniform disclosure form at www.icmje.org/coi_disclosure.pdf and declare: funding support from the Department of Dermatology, Northwestern University; no support from any organization for the submitted work; no financial relationships with any organizations that might have an interest in the submitted work in the previous three years; no other relationships or activities that could appear to have influenced the submitted work.

The lead author (MA) affirms that the manuscript is an honest, accurate, and transparent account of the study being reported; that no important aspects of the study have been omitted; and that any discrepancies from the study as planned (and, if relevant, registered) have been explained.

Dissemination to participants and related patient and public communities: We plan to share our work with readers by issuing a press release at the time of publication through Northwestern University’s media relations; tweeting about our work concurrently through our lab Twitter account (@AlamLab); distributing a plain language summary to regulators, advocacy groups, and clinicians; and writing an editorial with other physicians and patients in another journal that references this work.

Provenance and peer review: Not commissioned; externally peer reviewed.

Ethics statements

Ethical approval

Not required.

Data availability statement

All data relevant to the study can be requested from the corresponding author (m-alam@northwestern.edu).

References

  • 1.History of The Joint Commission. https://www.jointcommission.org/about-us/facts-about-the-joint-commission/history-of-the-joint-commission (accessed 2 Jun 2021).
  • 2.Mullner RM. Joint Commission. In: Encyclopedia of Health Services Research. Thousand Oaks: SAGE Publications. 2009. 663–8. 10.4135/9781412971942 [DOI] [Google Scholar]
  • 3.Joint Commission FAQs. https://www.jointcommission.org/about-us/facts-about-the-joint-commission/joint-commission-faqs (accessed 3 Jun 2021).
  • 4.Why Achieve Accreditation. https://www.jointcommission.org/accreditation-and-certification/become-accredited/why-achieve-accreditation (accessed 3 Jun 2021).
  • 5.State Recognition. https://www.jointcommission.org/accreditation-and-certification/state-recognition (accessed 3 Jun 2021).
  • 6.Institute of Medicine (US) Committee on Standards for Developing Trustworthy Clinical Practice Guidelines. Clinical Practice Guidelines We Can Trust. Washington (DC): National Academies Press (US) 2011. https://www.ncbi.nlm.nih.gov/books/NBK209539/ (accessed 5 Nov 2021).
  • 7.Standards for Guideline Development | United States Preventive Services Taskforce. https://www.uspreventiveservicestaskforce.org/uspstf/about-uspstf/methods-and-processes/standards-guideline-development (accessed 5 Nov 2021).
  • 8.The Joint Commission. Benefits of Joint Commission accreditation. 2021.https://www.jointcommission.org/-/media/tjc/documents/fact-sheets/benefits-of-joint-commission-accreditation-1-19-21.pdf
  • 9.About Our Standards. https://www.jointcommission.org/standards/about-our-standards (accessed 3 Jun 2021).
  • 10.R3 Report Issue 17 Distinct newborn identification requirement. https://www.jointcommission.org/standards/r3-report/r3-report-issue-17-distinct-newborn-identification-requirement/#.YpUQuJPMLeo Report Issue 17 Distinct newborn identification requirement (accessed 3 Jun 2021).
  • 11.R3 Report Issue 14 Pain assessment and management standards for ambulatory care. https://www.jointcommission.org/standards/r3-report/R3 Report Issue 14 Pain assessment and management standards for ambulatory care (accessed 3 Jun 2021).
  • 12.R3 Report Issue 15 Pain assessment and management standards for critical access hospitals. https://www.jointcommission.org/standards/r3-report/R3 Report Issue 15 Pain assessment and management standards for critical access hospitals (accessed 3 Jun 2021).
  • 13.R3 Report Issue 12 Maternal infectious disease status assessment and documentation standards for hos. https://www.jointcommission.org/standards/r3-report/R3 Report Issue 12 Maternal infectious disease status assessment and documentation standards for hos (accessed 3 Jun 2021).
  • 14.The Joint Commission. Accreditation Process Overview. 2021.https://www.jointcommission.org/-/media/tjc/documents/fact-sheets/accreditation-process-overview-5-10-21.pdf
  • 15.Standards. https://www.jointcommission.org/standards (accessed 3 Jun 2021).
  • 16.R3 Report. https://www.jointcommission.org/standards/r3-report (accessed 3 Jun 2021).
  • 17.Oxford Centre for Evidence-Based Medicine. Levels of Evidence (March 2009). Centre for Evidence-Based Medicine (CEBM), University of Oxford. https://www.cebm.ox.ac.uk/resources/levels-of-evidence/oxford-centre-for-evidence-based-medicine-levels-of-evidence-march-2009 (accessed 3 Jun 2021).
  • 18. Adashek JJ, Kurzrock R. Balancing clinical evidence in the context of a pandemic. Nat Biotechnol 2021;39:270-4. 10.1038/s41587-021-00834-6.  [DOI] [PubMed] [Google Scholar]
  • 19. Guyatt GH, Oxman AD, Vist GE, et al. GRADE Working Group . GRADE: an emerging consensus on rating quality of evidence and strength of recommendations. BMJ 2008;336:924-6. 10.1136/bmj.39489.470347.AD.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Cohen J. A Coefficient of Agreement for Nominal Scales. Educ Psychol Meas 1960;20:37-46. 10.1177/001316446002000104. [DOI] [Google Scholar]
  • 21. Prielipp RC, Coursin DB. All That Glitters Is Not a Golden Recommendation. Anesth Analg 2015;121:727-33. 10.1213/ANE.0000000000000499.  [DOI] [PubMed] [Google Scholar]
  • 22. Prasad V, Vandross A, Toomey C, et al. A decade of reversal: an analysis of 146 contradicted medical practices. Mayo Clin Proc 2013;88:790-8. 10.1016/j.mayocp.2013.05.012.  [DOI] [PubMed] [Google Scholar]
  • 23. Prasad V, Cifu A. Medical Reversal: Why We Must Raise the Bar Before Adopting New Technologies. Yale J Biol Med 2011;84:471-8. [PMC free article] [PubMed] [Google Scholar]
  • 24. Wilson ML. Regulations, Standards, Guidelines, and Benchmarks: A Need for Evidence-Based Management. Am J Clin Pathol 2016;145:742-3. 10.1093/ajcp/aqw108.  [DOI] [PubMed] [Google Scholar]
  • 25. Morey TE, Sappenfield JW, Gravenstein N, Rice MJ. Joint Commission and Regulatory Fatigue/Weakness/Overabundance/Distraction: Clinical Context Matters. Anesth Analg 2015;121:394-6. 10.1213/ANE.0000000000000732.  [DOI] [PubMed] [Google Scholar]
  • 26. Lam MB, Figueroa JF, Feyman Y, Reimold KE, Orav EJ, Jha AK. Association between patient outcomes and accreditation in US hospitals: observational study. BMJ 2018;363:k4011. 10.1136/bmj.k4011.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Flodgren G, Gonçalves-Bradley DC, Pomey MP. External inspection of compliance with standards for improved healthcare outcomes. Cochrane Database Syst Rev 2016;12:CD008992. 10.1002/14651858.CD008992.pub3.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Araujo CAS, Siqueira MM, Malik AM. Hospital accreditation impact on healthcare quality dimensions: a systematic review. Int J Qual Health Care 2020;32:531-44. 10.1093/intqhc/mzaa090.  [DOI] [PubMed] [Google Scholar]
  • 29. Hussein M, Pavlova M, Ghalwash M, Groot W. The impact of hospital accreditation on the quality of healthcare: a systematic literature review. BMC Health Serv Res 2021;21:1057. 10.1186/s12913-021-07097-6.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Cassel CK, Guest JA. Choosing wisely: helping physicians and patients make smart decisions about their care. JAMA 2012;307:1801-2. 10.1001/jama.2012.476.  [DOI] [PubMed] [Google Scholar]
  • 31. Walls RM, Resnick J. The Joint Commission on Accreditation of Healthcare Organizations and Center for Medicare and Medicaid Services community-acquired pneumonia initiative: what went wrong? Ann Emerg Med 2005;46:409-11. 10.1016/j.annemergmed.2005.07.018.  [DOI] [PubMed] [Google Scholar]
  • 32. Baker DW. History of The Joint Commission’s Pain Standards: Lessons for Today’s Prescription Opioid Epidemic. JAMA 2017;317:1117-8. 10.1001/jama.2017.0935.  [DOI] [PubMed] [Google Scholar]
  • 33. Reed JE, Howe C, Doyle C, Bell D. Successful Healthcare Improvements From Translating Evidence in complex systems (SHIFT-Evidence): simple rules to guide practice and research. Int J Qual Health Care 2019;31:238-44. 10.1093/intqhc/mzy160.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Vincent C, Amalberti R. Strategies for Safety. In: Safer Healthcare. Springer International Publishing, 2016. 10.1007/978-3-319-25559-0_6 [DOI] [PubMed] [Google Scholar]
  • 35. Mumford V, Forde K, Greenfield D, Hinchcliff R, Braithwaite J. Health services accreditation: what is the evidence that the benefits justify the costs? Int J Qual Health Care 2013;25:606-20. 10.1093/intqhc/mzt059.  [DOI] [PubMed] [Google Scholar]
  • 36. Rockwell DA, Pelletier LR, Donnelly W. The cost of accreditation: one hospital’s experience. Hosp Community Psychiatry 1993;44:151-5. 10.1176/ps.44.2.151.  [DOI] [PubMed] [Google Scholar]
  • 37.ISQua Guidelines and Principles for the Development of Health and Social Care Standards 5th Edition v1.0 (1).pdf.
  • 38.Healthcare organisations accreditation programme in France. Haute Autorité de Santé. https://www.has-sante.fr/jcms/c_2044304/en/healthcare-organisations-accreditation-programme-in-france (accessed 22 Jun 2021).
  • 39.National Institute for Health Care and Excellence. Timeline for developing quality standards. Standards and Indicators. NICE. https://www.nice.org.uk/standards-and-indicators/timeline-developing-quality-standards (accessed 3 Jun 2021).
  • 40. ACOG Committee Opinion No . Committee opinion no: 635: Prenatal and perinatal human immunodeficiency virus testing: expanded recommendations. Obstet Gynecol 2015;125:1544-7. 10.1097/01.AOG.0000466370.86393.d2  [DOI] [PubMed] [Google Scholar]
  • 41.Riley SE, Stark A, American Academy of Pediatrics, & American College of Obstetricians and Gynecologists. Guidelines for perinatal care. 7th ed. Elk Grove Village, IL; Washington, DC 2012. [Google Scholar]
  • 42. Nesheim S, Taylor A, Lampe MA, et al. A framework for elimination of perinatal transmission of HIV in the United States. Pediatrics 2012;130:738-44. 10.1542/peds.2012-0194.  [DOI] [PubMed] [Google Scholar]
  • 43.Group B Strep Prevention Guidelines. CDC. 2021.https://www.cdc.gov/groupbstrep/guidelines/index.html (accessed 3 Jun 2021).
  • 44.HIV among Pregnant Women, Infants, and Children. Atlanta, GA: CDC, National Center for HIV/AIDS, Viral Hepatitis, STD, and TB Prevention. Division of HIV/AIDS Prevention, March 2017. (Accessed March 22, 2021.) https://www.cdc.gov/hiv/pdf/group/gender/pregnantwomen/cdc-hiv-pregnant-women.pdf.
  • 45. Verani JR, Spina NL, Lynfield R, et al. Early-onset group B streptococcal disease in the United States: potential for further reduction. Obstet Gynecol 2014;123:828-37. 10.1097/AOG.0000000000000163.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Fitz Harris LF, Taylor AW, Zhang F, et al. Factors associated with human immunodeficiency virus screening of women during pregnancy, labor and delivery, United States, 2005-2006. Matern Child Health J 2014;18:648-56. 10.1007/s10995-013-1289-7.  [DOI] [PubMed] [Google Scholar]
  • 47.Recommendations for Use of Antiretroviral Drugs in Pregnant HIV-1 Infected Women for Maternal Health and Interventions to Reduce Perinatal HIV Transmission in the United States. Bethesda, MD: NIH, Office of AIDS Research Advisory Council (OARAC). Last update 2014. https://aidsinfo.nih.gov/contentfiles/lvguidelines/PerinatalGL.pdf.
  • 48. Nesheim S, Harris LF, Lampe M. Elimination of perinatal HIV infection in the USA and other high-income countries: achievements and challenges. Curr Opin HIV AIDS 2013;8:447-56. 10.1097/COH.0b013e3283636ccb.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49. Trepka MJ, Mukherjee S, Beck-Sagué C, et al. Missed Opportunities for Preventing Perinatal Transmission of Human Immunodeficiency Virus, Florida, 2007-2014. South Med J 2017;110:116-28. 10.14423/SMJ.0000000000000609.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. National Academy of Medicine (U.S.) , ed. First, do no harm: marshaling clinician leadership to counter the opioid epidemic. National Academy of Medicine, 2017. [PubMed] [Google Scholar]
  • 51.The Health Research & Educational Trust of the American Hospital Association. AHRQ Safety Program for Ambulatory Surgery. Final Report. AHRQ Publication May 2017 No. 16(17)-0019-1-EF.
  • 52. Mostofian F, Ruban C, Simunovic N, Bhandari M. Changing physician behavior: what works? Am J Manag Care 2015;21:75-84. [PubMed] [Google Scholar]
  • 53. Alford DP, Zisblatt L, Ng P, et al. SCOPE of pain: An evaluation of an opioid risk evaluation and mitigation strategy continuing education program. Pain Med 2016;17:52-63. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Chou R, Gordon DB, de Leon-Casasola OA, et al. Management of Postoperative Pain: A Clinical Practice Guideline From the American Pain Society, the American Society of Regional Anesthesia and Pain Medicine, and the American Society of Anesthesiologists’ Committee on Regional Anesthesia, Executive Committee, and Administrative Council. J Pain 2016;17:131-57. 10.1016/j.jpain.2015.12.008.  [DOI] [PubMed] [Google Scholar]
  • 55. Vadivelu N, Kai AM, Kodumudi V, Berger JM. Challenges of pain control and the role of the ambulatory pain specialist in the outpatient surgery setting. J Pain Res 2016;9:425-35. 10.2147/JPR.S86579.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Substance Abuse and Mental Health Services Administration. Prescription Drug Monitoring Programs: A Guide for Healthcare Providers. 2017.
  • 57.The Pew Charitable Trusts. Prescription drug monitoring programs: Evidence-based practices to optimize prescriber use 2016. 2016.
  • 58. Miller RM, Kaiser RS. Psychological Characteristics of Chronic Pain: a Review of Current Evidence and Assessment Tools to Enhance Treatment. Curr Pain Headache Rep 2018;22:22. 10.1007/s11916-018-0663-y.  [DOI] [PubMed] [Google Scholar]
  • 59. Reid MC, Eccleston C, Pillemer K. Management of chronic pain in older adults. BMJ 2015;350:h532. 10.1136/bmj.h532.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60. Stanos S, Brodsky M, Argoff C, et al. Rethinking chronic pain in a primary care setting. Postgrad Med 2016;128:502-15. 10.1080/00325481.2016.1188319.  [DOI] [PubMed] [Google Scholar]
  • 61. Tobias JD. Acute pain management in infants and children-Part 1: Pain pathways, pain assessment, and outpatient pain management. Pediatr Ann 2014;43:e163-8. 10.3928/00904481-20140619-10.  [DOI] [PubMed] [Google Scholar]
  • 62.Czarnecki M, Turner H. Core Curriculum for Pain Management Nursing. 3rd ed. Elsevier 2018.
  • 63. Becker WC, Edmond SN, Cervone DJ, et al. Evaluation of an Integrated, Multidisciplinary Program to Address Unsafe Use of Opioids Prescribed for Pain. Pain Med 2018;19:1419-24. 10.1093/pm/pnx041.  [DOI] [PubMed] [Google Scholar]
  • 64. Palermo T, et al. Assessment and management of children with chronic pain. American Pain Society, 2012: 4. [Google Scholar]
  • 65. Gordon DB, de Leon-Casasola OA, Wu CL, Sluka KA, Brennan TJ, Chou R. Research Gaps in Practice Guidelines for Acute Postoperative Pain Management in Adults: Findings From a Review of the Evidence for an American Pain Society Clinical Practice Guideline. J Pain 2016;17:158-66. 10.1016/j.jpain.2015.10.023.  [DOI] [PubMed] [Google Scholar]
  • 66. Cornelius R, Herr KA, Gordon DB, Kretzer K, Butcher HK. Evidence-Based Practice Guideline: Acute Pain Management in Older Adults. J Gerontol Nurs 2017;43:18-27. 10.3928/00989134-20170111-08.  [DOI] [PubMed] [Google Scholar]
  • 67. O’Brien EM, Staud RM, Hassinger AD, et al. Patient-centered perspective on treatment outcomes in chronic pain. Pain Med 2010;11:6-15. 10.1111/j.1526-4637.2009.00685.x.  [DOI] [PubMed] [Google Scholar]
  • 68. Mohammed K, Nolan MB, Rajjo T, et al. Creating a Patient-Centered Health Care Delivery System: A Systematic Review of Health Care Quality From the Patient Perspective. Am J Med Qual 2016;31:12-21. 10.1177/1062860614545124.  [DOI] [PubMed] [Google Scholar]
  • 69. Midboe AM, Lewis ET, Paik MC, et al. Measurement of adherence to clinical practice guidelines for opioid therapy for chronic pain. Transl Behav Med 2012;2:57-64. 10.1007/s13142-011-0104-5.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70. Parchman ML, Von Korff M, Baldwin L-M, et al. Primary Care Clinic Re-Design for Prescription Opioid Management. J Am Board Fam Med 2017;30:44-51. 10.3122/jabfm.2017.01.160183.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71. Adelman J, Aschner J, Schechter C, et al. Use of Temporary Names for Newborns and Associated Risks. Pediatrics 2015;136:327-33. 10.1542/peds.2015-0007.  [DOI] [PubMed] [Google Scholar]
  • 72.Adelman J, et al. The ‘Babyboy/Babygirl’ Problem: Evaluating the risk of non-distinct temporary first names for newborns and measuring the effect of changing the paradigm to reduce wrong patient orders. Paper presented at the Perinatal Section Opening Poster Session, San Diego, CA. October 2014. [Google Scholar]
  • 73.Patient ID a Top Source of Error; Newborns High Risk. https://www.reliasmedia.com/articles/139619-patient-id-a-top-source-of-error-newborns-high-risk?v=preview (accessed 4 Jun 2021).
  • 74. Gray JE, Suresh G, Ursprung R, et al. Patient misidentification in the neonatal intensive care unit: quantification of risk. Pediatrics 2006;117:e43-7. 10.1542/peds.2005-0291.  [DOI] [PubMed] [Google Scholar]
  • 75. The Joint Commission . Temporary names put newborns at risk. Quick Safety 2015;(17):1-2. [Google Scholar]
  • 76. Wallace SC. Newborns Pose Unique Identification Challenges. PA Patient Saf Advis 2016;13:9. [Google Scholar]
  • 77. Qaseem A, Wilt TJ, McLean RM, et al. Clinical Guidelines Committee of the American College of Physicians . Noninvasive Treatments for Acute, Subacute, and Chronic Low Back Pain: A Clinical Practice Guideline From the American College of Physicians. Ann Intern Med 2017;166:514-30. 10.7326/M16-2367.  [DOI] [PubMed] [Google Scholar]
  • 78. Skelly AC, Chou R, Dettori JR, et al. Noninvasive Nonpharmacological Treatment for Chronic Pain: A Systematic Review. Agency for Healthcare Research and Quality. AHRQ, 2018, 10.23970/AHRQEPCCER209. [DOI] [PubMed] [Google Scholar]
  • 79. Nahin RL, Boineau R, Khalsa PS, Stussman BJ, Weber WJ. Evidence-Based Evaluation of Complementary Health Approaches for Pain Management in the United States. Mayo Clin Proc 2016;91:1292-306. 10.1016/j.mayocp.2016.06.007.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80. Tick H, Nielsen A, Pelletier KR, et al. Pain Task Force of the Academic Consortium for Integrative Medicine and Health . Evidence-Based Nonpharmacologic Strategies for Comprehensive Pain Care: The Consortium Pain Task Force White Paper. Explore (NY) 2018;14:177-211. 10.1016/j.explore.2018.02.001.  [DOI] [PubMed] [Google Scholar]
  • 81. Dowell D, Haegerich TM, Chou R. CDC Guideline for Prescribing Opioids for Chronic Pain - United States, 2016. MMWR Recomm Rep 2016;65:1-49. 10.15585/mmwr.rr6501e1.  [DOI] [PubMed] [Google Scholar]
  • 82. Chung F, Memtsoudis SG, Ramachandran SK, et al. Society of Anesthesia and Sleep Medicine Guidelines on Preoperative Screening and Assessment of Adult Patients With Obstructive Sleep Apnea. Anesth Analg 2016;123:452-73. 10.1213/ANE.0000000000001416.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83. Lam KK, Kunder S, Wong J, Doufas AG, Chung F. Obstructive sleep apnea, pain, and opioids: is the riddle solved? Curr Opin Anaesthesiol 2016;29:134-40. 10.1097/ACO.0000000000000265.  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84. Lam T, Nagappa M, Wong J, Singh M, Wong D, Chung F. Continuous Pulse Oximetry and Capnography Monitoring for Postoperative Respiratory Depression and Adverse Events: A Systematic Review and Meta-analysis. Anesth Analg 2017;125:2019-29. 10.1213/ANE.0000000000002557.  [DOI] [PubMed] [Google Scholar]
  • 85.National Quality Partner Playbook. Opioid Stewardship. 2018.https://store.qualityforum.org/products/national-quality-partners-playbook%E2%84%A2-opioid-stewardship
  • 86. Walters TL, Howard SK, Kou A, et al. Design and Implementation of a Perioperative Surgical Home at a Veterans Affairs Hospital. Semin Cardiothorac Vasc Anesth 2016;20:133-40. 10.1177/1089253215607066.  [DOI] [PubMed] [Google Scholar]
  • 87. Alvis BD, King AB, Pandharipande PP, et al. Creation and Execution of a Novel Anesthesia Perioperative Care Service at a Veterans Affairs Hospital. Anesth Analg 2017;125:1526-31. 10.1213/ANE.0000000000001930.  [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary information: Supplementary tables 1 and 2

ibrs063064.ww.pdf (119.1KB, pdf)

Data Availability Statement

All data relevant to the study can be requested from the corresponding author (m-alam@northwestern.edu).


Articles from The BMJ are provided here courtesy of BMJ Publishing Group

RESOURCES