Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2023 Dec 1.
Published in final edited form as: J Patient Saf. 2022 Jul 21;18(8):e1231–e1236. doi: 10.1097/PTS.0000000000001062

Hospitals That Report Severe Sepsis and Septic Shock Bundle (SEP-1) Compliance Have More Structured Sepsis Performance Improvement

Ty B Bolte 1, Morgan B Swanson 1, Anna M Kaldjian 1, Nicholas M Mohr 1,2, Jennifer McDanel 3, Azeemuddin Ahmed 1
PMCID: PMC9722504  NIHMSID: NIHMS1818238  PMID: 35858483

Abstract

Purpose:

Sepsis is a common cause of death. The CMS SEP-1 bundle is focused on improving sepsis outcomes, but it is unknown which quality improvement (QI) practices are associated with SEP-1 compliance and reduced sepsis mortality. The objectives of this study were to compare sepsis QI practices in SEP-1 reporting and non-reporting hospitals, and to measure the association between sepsis QI processes, SEP-1 performance, and sepsis mortality.

Materials and Methods:

This study linked survey data on QI practices from Iowa hospitals to SEP-1 performance data and mortality. Characteristics of hospitals and sepsis QI practices were compared by SEP-1 reporting status. Univariable and multivariable logistic and linear regression estimated the association of QI practices with SEP-1 performance and observed-to-expected sepsis mortality ratios.

Results:

100% of Iowa’s 118 hospitals completed the survey. SEP-1 reporting hospitals were more likely to have sepsis QI practices, including reporting sepsis quality to providers (64% vs. 38%, p=0.026) and using the case review process to develop sepsis care plans (87% vs. 64%, p=0.013). Sepsis QI practices were not associated with increased SEP-1 scores. A sepsis registry was associated with decreased odds of being in the bottom quartile of sepsis mortality (OR: 0.37; 95%CI 0.14–0.96, p=0.041), and presence of a sepsis committee was associated with lower hospital-specific mortality (O:E ratio: −0.11; 95%CI −0.20--0.01).

Conclusions:

Hospitals reporting SEP-1 compliance conduct more sepsis QI practices. Most QI practices are not associated with increased SEP-1 performance or decreased sepsis mortality. Future work could explore how to implement these performance improvement practices in hospitals not reporting SEP-1 compliance.

Keywords: SEP-1, sepsis, quality improvement, rural

Introduction:

Each year, approximately 1.7 million adults are diagnosed with sepsis, resulting in more than 250,000 deaths. Health care costs in the U.S. due to sepsis are estimated to be over $62 billion annually. (1), (2) Performance improvement activities have been a hallmark of sepsis care since the Surviving Sepsis Campaign developed “sepsis bundles”, protocols detailing care for sepsis patients, and multidisciplinary quality improvement programs (3). In addition, performance improvement programs aimed at improving sepsis bundle treatment have demonstrated marked reduction in sepsis mortality even after adjustment for comorbidities, age, and disease severity. (4) Other studies have confirmed the benefits of using sepsis bundles to improve patient outcomes.(3-6) Based on this work, the Centers for Medicare & Medicaid Services (CMS) adopted the Early Management Bundle, Severe Sepsis/Septic Shock (SEP-1) quality measure in 2015.(6, 7, 8, 9, 10) The SEP-1 quality measure is a process measure that scores hospitals on compliance with the SEP-1 sepsis bundle, and provides a score. SEP-1 scores are a percentage from 0 to 100 and represent the proportion of eligible sepsis cases with complete adherence with both [1]) a 3-hour bundle of serum lactate measurement, fluid resuscitation, blood cultures, and broad-spectrum antibiotics and [2] a 6-hour bundle including repeat lactate measurement, vasopressor administration, and reassessment of volume status/tissue perfusion. Additionally, failure of one aspect of the 7 bundle elements will result in failure of the entire SEP-1 measure. It is also important to differentiate a process measure such as SEP-1 which looks at a volume at which a practice is being performed versus an outcome measure in which a process is compared to outcomes. Since its introduction, the SEP-1 measure has drawn attention to sepsis quality and focused the efforts of hospital quality departments on sepsis performance. (11)

CMS measures provide process metrics that are commonly used for quality measurement and improvement in a variety of conditions, including heart failure, osteoporosis, stroke, and myocardial infarction.(12-16) Many hospitals have used these initiatives to narrow the focus of disease-specific quality initiatives, but a recent survey of hospital quality officers found that the complexity of SEP-1 was an obstacle in meeting the requirements of this measure.(17) These findings make SEP-1 an important focus not only for improving sepsis outcomes, but also for understanding how quality measurement impacts the structure and process of performance improvement. Additionally, low volume hospitals (five cases of sepsis per quarter or less) are not required to report SEP-1 performance to CMS, and thus many are not following the bundle.

The objectives of this study are to [1] compare sepsis performance improvement activities in hospitals that do and do not report SEP-1 performance and [2] to identify elements of sepsis performance improvement associated with better SEP-1 scores and outcomes. Our goal is to identify elements of sepsis performance improvement processes associated with better SEP-1 scores and outcomes to inform hospital performance improvement programs.

Methods:

Study Setting and Design.

This study was a mixed methods analysis linking hospital performance improvement activities measured from a telephone survey of hospital sepsis coordinators with CMS SEP-1 reporting data and clinical outcomes measured using all-payer administrative claims in a rural, Midwestern state. Our survey of quality improvement and safety coordinators at all acute care hospitals in the state was completed between May 2020 to July 2020. Outcome data for sepsis bundle adherence and hospital aggregate sepsis mortality were linked from publicly available hospital quality reporting data and risk-adjusted mortality calculated from statewide all-payer inpatient and ED administrative claims data. This study was determined not to be human subjects research by the local Institutional Review Board and is reported according to the STrengthening the Reporting of OBservational studies in Epidemiology (STROBE) guidelines.(18)

Survey of Hospital Performance Improvement Practices.

A telephone survey of hospital quality improvement and safety coordinators was conducted to determine the presence and characteristics of sepsis performance improvement practices in acute care hospitals. The sampling frame was all non-federal acute care hospitals in the state (n=118 hospitals). The questionnaire (Supplement Appendix 1) was developed and refined by the study team, including an ED physician-administrator (AA), a quality engineer with a doctoral degree in epidemiology (JM), a critical care physician and health services researcher (NM), and research staff conducting surveys (TB, MB, AK). The instrument consisted of both structured questions and semi-structured prompts, intended to collect diverse responses representing the breadth of health care delivery, without restricting our data to findings expected by the study team. The questionnaire was field tested in two institutions for face validity and flow. Two trained research team members conducted the survey (TB and AK), and standardized prompts and a survey script were used to ensure uniformity across respondents. The survey questionnaire was designed to assess for the presence and characteristics of performance improvement practices in eight categories: sepsis committee, sepsis coordinator, physician sepsis champion, sepsis case review process, code sepsis response team (a team activated to provide care when a sepsis patient is identified) , standardized process for sepsis patient identification, sepsis training, and sepsis registry. Survey respondents were identified by contacting hospitals via telephone and asking for the Director of Quality Improvement, if the hospital did not have this position, someone with a similar role was identified via speaking with hospital staff. If a respondent did not know the answer to a particular survey question, follow-up calls were scheduled with other members of the hospital team. Definitions used for the survey process are included in the Supplement Appendix 2. Data were collected using Research Electronic Data Capture software (REDCap) (Vanderbilt University, Nashville, TN).

Data Sources and Definitions.

Survey data were linked to two data sources: CMS SEP-1 score data and statewide hospital and hospital administrative claims data. The most recent SEP-1 score data were obtained from the CMS Hospital Compare public website for the reporting periods of April 2018 to March 2019.(19) State hospital association data containing administrative claims for all adults (18 years) treated in a hospital within the state between 2015 to 2019 were used to estimate sepsis risk-adjusted mortality by hospital.(20) Both datasets were deterministically linked using a unique hospital identifier to the telephone survey data. Hospital characteristics additionally included annual sepsis volume (from claims data), critical access hospital status (from CMS data, federally qualified small rural hospitals that are structurally different from higher volume centers), and rurality of hospital (from the zip-code tabulation area approximation of Rural-Urban Commuting Areas [RUCA] codes).(21) It is important to note that SEP-1 reporting is absolute meaning that a case either passes by meeting all criteria, or fails, and reporting is guided by whether there is a mandate for a hospital to report, and hospitals with less than five cases of sepsis per quarter are encouraged but not required to submit patient level data (19).

Key Measures & Outcomes.

The primary exposure in this analysis was sepsis performance improvement activities, and the primary outcome in this study was hospital SEP-1 reporting status. Hospital SEP-1 reporting status was defined by presence of any SEP-1 scores during the reporting periods. Performance improvement measures were obtained from the study survey (Appendix 1). These measures were evaluated as binary (present/absent) for the primary analysis with characteristics of the components assessed in secondary analyses.

Secondary outcomes included SEP-1 adherence and hospital-specific risk-adjusted mortality. SEP-1 scores were retained as both a continuous measure and categorized by quartile (e.g., top quartile is 75th-99th percentile) of hospitals.

Risk-adjusted sepsis mortality was defined as hospital-specific observed-to-expected (O:E) sepsis mortality, assigning transferred cases to the first hospital. All adult patients presenting to the ED or inpatient unit with sepsis, fed by the implicit definition combining at least one infection diagnosis and one organ failure/dysfunction diagnosis code, were included in the mortality calculation.(20, 22) Expected mortality was calculated by estimating predicted probabilities for mortality for each case using a multivariable logistic regression model (age, race, sex, year, Elixhauser comorbidities (23), patient rurality, infection source, sepsis ICD-10-CM codes, palliative care, and organ dysfunction at admission) to generate a predicted probability of in-hospital mortality. The O:E ratio was then calculated as the ratio of the sum of observed mortality divided by the sum of the expected probability of mortality for each hospital.

Data Analysis.

Characteristics of hospitals reporting SEP-1 data were compared to hospitals not reporting SEP-1 data using descriptive statistics and chi-square tests. Chi-squared and Fisher’s exact tests were used to compare characteristics of sepsis performance improvement practices by SEP-1 reporting status. SEP-1 scores and O:E mortality ratios were compared by sepsis performance improvement practices with Wilcoxon Mann Whitney test and across quartiles (gamma). Logistic and linear regression were used to identify associations between sepsis performance improvement practices and top- and bottom- preforming hospitals (defined as top and bottom 25% of SEP-1 hospitals, respectively) and SEP-1 scores, respectively. Quartiles were used to screen for threshold effects for high-performing hospitals, and then the continuous outcome was used to assess for associations between performance improvement practices and SEP-1 scores. For linear regression, the assumptions of the exposure-outcome relationship and homoskedasticity of residuals were assessed visualizing the pattern of observed versus predicted residuals. For the secondary outcome of O:E mortality ratios, a log-transformed model was used to satisfy these assumptions. Models adjusting for rurality of the hospital and critical access status were considered, but when relationships between rurality and critical access status and the outcome were not observed, only unadjusted models were presented. Statistical tests were considered significant at the p<0.05 threshold for two-tailed tests; analyses were conducted in SAS (version 9.4; SAS Institute, Cary, NC).

Post-Hoc Sensitivity and Stratified Analyses.

A stratified analysis assessed for differences in the relationship between performance improvement practices and sepsis mortality by SEP-1 reporting status. Two subcomponents (case review results reported to provider and results used in sepsis care plans) were hypothesized to be independently related with outcomes, so regression models estimated the association of these specific practices with outcomes. Last, a sensitivity analysis was performed assigning hospital outcomes based on last hospital for transferred patients (rather than first hospital).(24)

Results:

Description of Study Population.

All hospitals responded to the telephone survey (response rate 100%, N=118). Overall, hospitals had a median sepsis volume of 121 sepsis patients per year, 22% of hospitals were urban and 70% were rural critical access hospitals (Table 1). Forty-four hospitals (37%) reported SEP-1 adherence and the median SEP-1 score was 56.5 (IQR: 44.5 to 71.5). SEP-1 reporting hospitals were more likely to be urban (48% vs. 7%; %diff: 41%, 95%CI: 24 to 56%) and less likely to be critical access hospitals (23% vs. 97%; %diff: −74%, 95%CI: −84 to −58%) when compared to hospitals that did not report SEP-1 metrics. Top-performing hospitals (i.e., those in the top quartile) had SEP-1 scores between 73 and 97.

Table 1.

Hospital characteristics by SEP-1 reporting status.

Hospital Characteristic Overall
(N=118)
SEP-1 Reporting
Hospitals (n=44)
Non-SEP-1 Reporting
Hospitals (n=74)
p-value
N or
Median
% or IQR n or
Median
% or IQR n or
Median
% or IQR
Sepsis Volume (Sepsis Visits per Year) 121 68 – 326 933 244 – 1,494 78 56 - 697 <0.001
Rurality (RUCA) <0.001
 Urban 26 22.0 21 47.7 5 6.8
 Large Rural City/Town 15 12.7 13 29.6 2 2.7
 Small Rural City/Town 50 42.4 8 18.2 42 56.8
 Isolated Small Rural Town 27 22.9 2 4.6 25 33.8
Critical Access Hospital Status 82 69.5 10 22.7 72 97.3 <0.001
SEP-1 Score (Median, IQR) - - 56.5 44.5 – 71.5 - - -

IQR=Interquartile range; RUCA= Rural-Urban Commuting Area

Comparison of Performance Improvement Practices by SEP-1 Reporting Status.

Overall, the majority of hospitals had a formal sepsis case identification process (92%, n=109), sepsis case review (75%, n=89), and sepsis training (66%, n=77). Other sepsis performance improvement practices were less common, including having a sepsis committee (39%, n=46) and a code sepsis response team (13%, n=15). Seven of the eight sepsis performance improvement components: sepsis committee (73 vs. 19%; difference 54%, 95%CI: 36 to 67%), sepsis coordinator (61 vs. 37%; difference 24%, 95%CI: 5 to 40%), physician sepsis champion (66 vs. 31%; difference 35%, 95%CI: 16 to 50%), sepsis case review process (89 vs. 68%; difference 21%, 95%CI: 5 to 34%), code sepsis response team (23 vs. 7%, ; difference 16%, 95%CI: 3 to 31%), sepsis training (89 vs. 51%; difference 38%, 95%CI: 21 to 51%), and sepsis registry (66 vs. 26%; difference 40%, 95%CI: 22 to 55%) were more commonly reported in SEP-1 reporting hospitals (Figure 1). There was no difference in the use of a standardized process for sepsis patient identification between SEP-1 reporting and non-reporting hospitals.

Figure 1.

Figure 1.

Sepsis performance improvement practices by SEP-1 reporting status.

Among hospitals that reported each sepsis performance improvement practice, we assessed differences in the characteristics of how the performance improvement practice was performed (Table S1). Overall, most hospitals performed the improvement practices similarly. For example, the 46 hospitals with sepsis committees, most practices were similar except more SEP-1 hospitals reported decreasing sepsis mortality as a committee goal and tracked sepsis mortality and bundle adherence. Sepsis committees were commonly chaired by clinicians or nurses with members including physicians, hospital administrators, pharmacists, nurses, and quality/safety specialists; most met monthly or quarterly and tracked sepsis mortality and sepsis bundle adherence. Additionally, among the hospitals that had Code Sepsis response teams, there were no differences in the components of the team; with teams being staffed 24/7 with physicians and nurses, activated by SIRS criteria or clinician concern, and being utilized hospital-wide.

However, there were differences in practices by SEP-1 reporting status for sepsis coordinators, physician sepsis champions, sepsis case review, sepsis case identification, sepsis training, and sepsis registry (Figure 2A-B)(Table S1).

Figure 2.

Figure 2.

Sepsis performance improvement sub-components that differ by SEP-1 reporting status.

Association of Performance Improvement Practices with SEP-1 Scores.

There was no association observed between sepsis performance improvement practices and low-performance (i.e., bottom-quartile) SEP-1 hospitals (Figure 3) (Table S2). There remained no association when the SEP-1 outcome was redefined as a continuous outcome or high-performance (i.e., top-quartile) SEP-1 hospitals.

Figure 3. Association of sepsis performance improvement with SEP-1 scores and O:E mortality ratios (predicting bottom quartile performance).

Figure 3.

*Since all top-performing hospitals and most SEP-1 reporting hospitals had standardized processes for sepsis identification and sepsis training, the odds of top-quartile performance improvement practices could not be estimated.

Association of Performance Improvement Practices with Mortality.

Presence of a sepsis registry was associated with decreased odds of being in the bottom quartile of sepsis mortality (OR: 0.37; 95%CI 0.14 – 0.96, p=0.041) (Figure 3) (Table S3). When mortality was analyzed as a continuous outcome or as high-performance hospitals, these effects were no longer observed (Table S3). In this continuous model, presence of a sepsis committee was associated with lower hospital-specific mortality (log-transformed O:E ratio: −0.11; 95%CI −0.20 to −0.01).

Association of Subcomponents with SEP-1 scores and Mortality.

Two subcomponents of the case review process were independently assessed for associations with sepsis outcomes. Case review results reported by providers was not associated with SEP-1 scores (−0.43, 95%CI −2.75 – 1.88, p=0.708) or mortality (−0.03, 95%CI −0.13 – 0.07, p=0.585). Case review results used for sepsis care plans was not associated with SEP-1 scores (−1.29, 95%CI −4.00 – 1.42, p=0.342), but was associated with improved mortality (−0.10, 95%CI −0.19 to −0.00, p = 0.049).

Sensitivity Analyses.

When we reallocated transferred sepsis cases to the last hospital where they were treated (rather than the first), the results were similar (Table S4).

Discussion:

In our survey of hospitals in a predominantly rural state, we found that hospitals that report SEP-1 adherence were more likely to use performance improvement practices than those hospitals that do not report SEP-1. Among those that reported SEP-1 adherence, we did not find an association between individual performance improvement practices and SEP-1 score, but we did find that use of a sepsis registry, having a multidisciplinary sepsis committee, and using results from individual sepsis case review to refine sepsis protocols were all associated with decreased probability of a hospital being in the bottom quartile of sepsis risk-adjusted mortality. Perhaps most interesting, we observed that sepsis performance improvement practices are heterogeneous across hospitals, with very different local performance improvement programs.

One apparent explanation for this relationship could be that increased regulatory pressure facing hospitals who report SEP-1 may drive commitment to sepsis quality improvement. Seven of the eight sepsis performance improvement components were more commonly reported in SEP-1 reporting hospitals, and these hospitals also used their sepsis performance data for more quality improvement activities. Additionally, the rigid nature of the SEP-1 performance metric most likely contributed to the failure to detect better scores among facilities with more improvement activities, as failure of one aspect of the 7 bundle elements will result in failure of the entire SEP-1 measure.

It is important to note that these results do not indicate practices which were not corelated with improved outcomes are ineffective, it simply means they did not distinguish between high and low performing hospitals. Stated another way, it is likely these practices have a positive effect to sepsis care, but if every hospital is doing them, they will not be associated with improved outcomes. This means these practices are being utilized effectively and should be continued. The practices that were found to be associated with improved performance, however, should be expanded.

A relevant factor in our study is the types of hospitals included in the state where the study was conducted. As previously mentioned, SEP-1 reporting is not required of low-volume hospitals, and many hospitals that do not report SEP-1 are critical access hospitals. This factor alone might influence the sepsis quality improvement because these hospitals have fewer resources to focus on sepsis quality, lower sepsis volume, and many sepsis patients are transferred to other hospitals for inpatient care (23).

The relationship between performance improvement and mortality, however, suggests that there may be an important relationship between performance improvement and patient outcomes. The three factors associated with mortality all are related to the use of data for process improvement: having a sepsis registry, using data from the registry to improve protocols, and having a multidisciplinary sepsis committee. This observation is especially interesting because of the difficulty in collecting and maintaining accurate sepsis quality data—the SEP-1 measure is extremely complex and requires manual chart review. Diagnosis codes are imprecise, (24) timing of care is complicated (25), and the roster of clinicians responsible for providing sepsis care is large and heterogeneous.

This finding also highlights another important issue—sepsis quality measures exempt low-volume hospitals. While this is important from the perspective of measurement, it is also important if a lack of quality measurement leads to insufficient incentive for performance improvement. An important next step is to consider ways to include low-volume hospitals in sepsis performance measurement, as prior work has shown performance improvement yielded larger returns in urban vs. rural hospitals.(26) Previous studies have found that rural hospitals that report SEP-1 perform well, but any metric of rural sepsis quality is hampered by low rural hospital reporting.(27) Because quality reporting is required for prospective payment system (PPS) hospitals only and does not include critical access hospitals, a significant change in reporting requirements would require a change in policy.This study has several limitations. To address the impact of acute changes due to the COVID-19 pandemic, hospitals were asked to answer based on their practices prior to the pandemic (i.e., December 2019). All survey data were collected from May 2020 to July 2020, which meant any changes were recent enough that respondents were acutely aware of changes due to Covid. Thus, any differences were apparent and recall bias should be minimal. Another limitation was the temporal impact of changes over time on SEP-1 scores and mortality. However, the primary outcome (SEP-1 scores) occurred only 15 months before the survey, so we do not expect significant directional bias. Another limitation is the observational nature of our data collection, which limits our ability to draw causal conclusions. The ability to measure actual outcomes and heterogenous performance improvement practices over a series of health systems, though, mitigates that limitation.

Another limitation arises due to multiple comparisons; results were not adjusted for multiple comparisons because this study was hypothesis-generating and focused on identifying the strongest predictors of performance. The results of these analyses, however, should be considered in the context of the larger hospital network in which this study was conducted, and findings should be replicated by prospective interventional trials before we view the findings with confidence that intervention will change performance.

Although the models for critical access hospitals (CAH) and non-CAH were run separately, reporting status is nearly colinear with sepsis volume and CAH status. This is because critical access hospitals are not required to report SEP-1 performance, and therefore fell almost entirely in the non-SEP 1 reporting category. CAHs and small hospitals may be less likely to employ sepsis QI practices due to other reasons, including funding and staffing limitations. However, because the SEP-1 measure doesn’t apply to CAHs in general, non-reporting hospitals may still have good QI performance, as prior work has suggested (28, 29).

Finally, our sample was drawn from a single Midwestern state. Despite that, the generalizability of this study is broad: the state where the study was conducted reflects the national distribution of SEP-1 scores well; with a state average of 51% compared to the national average of 50% (30) and includes rural and urban hospitals with a range of sepsis volumes.

Conclusion:

In conclusion, hospitals that report SEP-1 performance are engaged in more performance improvement activities than those that do not. Some of these activities, including the presence of an institutional sepsis registry, using sepsis quality data to change hospital protocols, and having a multidisciplinary sepsis committee are associated with decreased hospital risk-adjusted mortality, but a relationship with SEP-1 performance was not observed. This finding is important in understanding the role of quality measurement on the process of performance improvement.

Future research should be focused on broadening the reach and scope of performance improvement practices. Public and government organizations may encourage or incentivize these activities, and these interventions may be associated with improved outcomes. Most important, future work should seek to validate which of these performance improvement activities are most effective in improving outcomes, because hospitals would value an evidence-based framework for performance improvement.

Supplementary Material

Supplemental Data File (.doc, .tif, pdf, etc.)

Funding:

Funding was supported by the National Institutes of Health (NIH) National Heart, Lung, and Blood Institute (T35 HL007485), the University of Iowa Institute for Clinical and Translational Sciences funded by a grant from the National Center for Advancing Translational Sciences (U54TR001356), and the University of Iowa Department of Emergency Medicine. Ms. Swanson is additionally supported by a grant from the National Center for Child Health and Human Development (F30HD100074) and Dr. Mohr is additionally supported by a grant from the Agency for Healthcare Research and Quality (K08 HS025753). These contents are solely the responsibility of the authors and do not necessarily reflect the views of the National Institutes of Health, the Agency for Healthcare Research and Quality, or the U.S. Department of Health and Human Services.

Footnotes

Conflicts of Interest: No conflicts of interest were declared.

Institution Where the Work was Performed: University of Iowa Carver College of Medicine, Iowa City, Iowa

References

  • 1.Rhee C, Jones TM, Hamad Y, Pande A, et al. : Prevalence, Underlying Causes, and Preventability of Sepsis-Associated Mortality in US Acute Care Hospitals. JAMA Netw Open 2019; 2(2):e187571. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Whitfield PL, Ratliff PD, Lockhart LL, Andrews D, et al. : Implementation of an adult code sepsis protocol and its impact on SEP-1 core measure perfect score attainment in the ED. Am J Emerg Med 2020; 38(5):879–882 [DOI] [PubMed] [Google Scholar]
  • 3.Levy MM, Dellinger RP, Townsend SR, Linde-Zwirble WT, et al. : The Surviving Sepsis Campaign: results of an international guideline-based performance improvement program targeting severe sepsis. Intensive Care Med 2010; 36(2):222–231 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Rivers E, Nguyen B, Havstad S, Ressler J, et al. : Early goal-directed therapy in the treatment of severe sepsis and septic shock. N Engl J Med 2001; 345(19):1368–1377 [DOI] [PubMed] [Google Scholar]
  • 5.Ferrer R, Artigas A, Levy MM, Blanco J, et al. : Improvement in process of care and outcome after a multicenter severe sepsis educational program in Spain. JAMA 2008; 299(19):2294–2303 [DOI] [PubMed] [Google Scholar]
  • 6.Walkey AJ, Lindenauer PK: Keeping It Simple in Sepsis Measures. J Hosp Med 2017; 12(12):1019–1020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Allison MG, Schenkel SM: SEP-1: A Sepsis Measure in Need of Resuscitation? Ann Emerg Med 2018; 71(1):18–20 [DOI] [PubMed] [Google Scholar]
  • 8.Gesten F, Evans L. SEP-1—Taking the Measure of a Measure. JAMA Netw Open. 2021;4(12):e2138823. doi: 10.1001/jamanetworkopen.2021.38823 [DOI] [PubMed] [Google Scholar]
  • 9.Rhee C, Yu T, Wang R, et al. Association Between Implementation of the Severe Sepsis and Septic Shock Early Management Bundle Performance Measure and Outcomes in Patients With Suspected Sepsis in US Hospitals. JAMA Netw Open. 2021;4(12):e2138596. doi: 10.1001/jamanetworkopen.2021.38596 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Winslow Rosalie, Failing the metric but saving lives: The protocolization of sepsis treatment through quality measurement, Social Science & Medicine, Volume 253, 2020, 112982, ISSN 0277–9536, 10.1016/j.socscimed.2020.112982. [DOI] [PubMed] [Google Scholar]
  • 11.Bauer SR, Han X, Wang XF, Blonsky H, et al. : Association Between Compliance With the Sepsis Quality Measure (SEP-1) and Hospital Readmission. Chest 2020; 158(2):608–611 [DOI] [PubMed] [Google Scholar]
  • 12.Hunt JR, Ouellette KJ, Reece M: Using Lean to Enhance Heart Failure Patient Identification Processes and Increase Core Measure Scores. J Nurs Care Qual 2019; 34(1):28–33 [DOI] [PubMed] [Google Scholar]
  • 13.French S, Choden S, Schmajuk G: Quality Measures and Quality Improvement Initiatives in Osteoporosis-an Update. Curr Osteoporos Rep 2019; 17(6):491–509 [DOI] [PubMed] [Google Scholar]
  • 14.Poisson SN, Josephson SA: Quality measures in stroke. Neurohospitalist 2011; 1(2):71–77 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Fonarow GC, Zhao X, Smith EE, Saver JL, et al. : Door-to-needle times for tissue plasminogen activator administration and clinical outcomes in acute ischemic stroke before and after a quality improvement initiative. JAMA 2014; 311(16):1632–1640 [DOI] [PubMed] [Google Scholar]
  • 16.Brackbill ML, Kline VT, Sytsma CS, Call JT: Intervention to increase the proportion of acute myocardial infarction or coronary artery bypass graft patients receiving an order for aspirin at hospital discharge. J Manag Care Pharm 2010; 16(5):329–336 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Barbash IJ, Rak KJ, Kuza CC, Kahn JM: Hospital Perceptions of Medicare's Sepsis Quality Reporting Initiative. J Hosp Med 2017; 12(12):963–968 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.von Elm E, Altman DG, Egger M, Pocock SJ, et al. : The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement: guidelines for reporting observational studies. J Clin Epidemiol 2008; 61(4):344–349 [DOI] [PubMed] [Google Scholar]
  • 19.U.S. Department of Health & Human Services. Centers for Medicare & Medicaid Services: Hospital Compare—A Quality Tool for Adults, Including People with Medicare Available at: http://www.medicare.gov/hospitalcompare/search.html.Accessed [Google Scholar]
  • 20.Rudd KE, Johnson SC, Agesa KM, Shackelford KA, et al. : Global, regional, and national sepsis incidence and mortality, 1990–2017: analysis for the Global Burden of Disease Study. Lancet 2020; 395(10219):200–211 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.U.S. Department of Agriculture ERS: ERS/USDA Briefing Room: Measuring Rurality: Rural-Urban Commuting Area Codes. Available at: http://www.ers.usda.gov/briefing/Rurality/RuralUrbanCommutingAreas/. Accessed 15 October 2017,
  • 22.Angus DC, Linde-Zwirble WT, Lidicker J, Clermont G, et al. : Epidemiology of severe sepsis in the United States: analysis of incidence, outcome, and associated costs of care. Crit Care Med 2001; 29(7):1303–1310 [DOI] [PubMed] [Google Scholar]
  • 23.Elixhauser A, Steiner C, Harris DR, Coffey RM: Comorbidity measures for use with administrative data. Med Care 1998; 36(1):8–27 [DOI] [PubMed] [Google Scholar]
  • 24.Jolley RJ, Sawka KJ, Yergens DW, Quan H, Jetté N, Doig CJ. Validity of administrative data in recording sepsis: a systematic review. Crit Care. 2015;19(1):139. Published 2015 Apr 6. doi: 10.1186/s13054-015-0847-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Rhee C, Brown SR, Jones TM, O'Brien C, Pande A, Hamad Y, Bulger AL, Tobin KA, Massaro AF, Anderson DJ, Warren DK, Klompas M; CDC Prevention Epicenters Program. Variability in determining sepsis time zero and bundle compliance rates for the centers for medicare and medicaid services SEP-1 measure. Infect Control Hosp Epidemiol. 2018. Aug;39(8):994–996. doi: 10.1017/ice.2018.134. Epub 2018 Jun 22. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Wray D, Coppin JD, Scott D, Jacob DA, et al. : Increased HCV Screening Yields Discordant Gains in Diagnoses Among Urban and Rural Veteran Populations in Texas: Results of a Statewide Quality Improvement Initiative. Jt Comm J Qual Patient Saf 2019; 45(2):112–122 [DOI] [PubMed] [Google Scholar]
  • 27.Greenwood-Ericksen MB, Rothenberg C, Mohr N, Andrea SD, et al. : Urban and Rural Emergency Department Performance on National Quality Metrics for Sepsis Care in the United States. J Rural Health 2019; 35(4):490–497 [DOI] [PubMed] [Google Scholar]
  • 28.Greenwood-Ericksen M, Kamdar N, Lin P, George N, Myaskovsky L, Crandall C, Mohr NM, Kocher KE. Association of Rural and Critical Access Hospital Status With Patient Outcomes After Emergency Department Visits Among Medicare Beneficiaries. JAMA Netw Open. 2021. Nov 1;4(11):e2134980. doi: 10.1001/jamanetworkopen.2021.34980. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Greenwood-Ericksen MB, Rothenberg C, Mohr N, Andrea SD, Slesinger T, Osborn T, Whittle J, Goyal P, Tarrant N, Schuur JD, Yealy DM, Venkatesh A. Urban and Rural Emergency Department Performance on National Quality Metrics for Sepsis Care in the United States. J Rural Health. 2019. Sep;35(4):490–497. doi: 10.1111/jrh.12339. Epub 2018 Nov 28. [DOI] [PubMed] [Google Scholar]
  • 30.Kempker JA, Kramer MR, Waller LA, Wang HE, Martin GS. State-level hospital compliance with and performance in the Centers for Medicaid & Medicare Services' Early Management Severe Sepsis and Septic Shock Bundle. Crit Care. 2019. Mar 18;23(1):92. doi: 10.1186/s13054-019-2382-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Mohr NM, Harland KK, Shane DM, Ahmed A, et al. : Inter-hospital transfer is associated with increased mortality and costs in severe sepsis and septic shock: An instrumental variables approach. J Crit Care 2016; 36:187–194 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Miller RR 3rd, Dong L, Nelson NC, Brown SM, Kuttler KG, Probst DR, Allen TL, Clemmer TP; Intermountain Healthcare Intensive Medicine Clinical Program. Multicenter implementation of a severe sepsis and septic shock treatment bundle. Am J Respir Crit Care Med. 2013. Jul 1;188(1):77–82. doi: 10.1164/rccm.201212-2199OC. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Barbash IJ, Davis B, Kahn JM. National Performance on the Medicare SEP-1 Sepsis Quality Measure. Crit Care Med. 2019. Aug;47(8):1026–1032. doi: 10.1097/CCM.0000000000003613. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplemental Data File (.doc, .tif, pdf, etc.)

RESOURCES