Skip to main content
Open Forum Infectious Diseases logoLink to Open Forum Infectious Diseases
. 2025 Oct 22;12(Suppl 2):S1391–S1403. doi: 10.1093/ofid/ofaf489

Beyond Accuracy: Methodological Advances for Assessing the Clinical Impact of Infectious Disease Diagnostics

Kimberly C Claeys 1, Andrea M Prinzi 2,✉,2, Tristan T Timbrook 3
PMCID: PMC12541908  PMID: 41132515

Abstract

Evaluating the clinical impact of in vitro diagnostic tests (IVDs) for infectious diseases is complex given their effectiveness depends on context, implementation, and provider behavior. Traditional methodologies for therapy interventions do not adequately capture this complexity, necessitating novel analytical approaches and study designs. This review highlights methodological considerations for improving evidence generation for infectious diseases IVDs. Design and analysis challenges leading to bias and related solutions are reviewed such as the target trial framework. Moreover, novel frameworks such as Benefit–Risk Evaluation of Diagnostics: A Framework, Desirability of Outcome Ranking Management of Antimicrobial Therapy, and Desirability of Outcome Ranking and study designs such as hybrid effectiveness–implementation designs are discussed which allow for holistic ways to assess real-world outcomes. By evaluating IVDs with practical, real-world evidence, tests can better inform clinical decision making, policy, and ultimately patient outcomes.

Keywords: diagnostics, hybrid effectiveness-implementation, infectious diseases


Assessing the clinical impact of in vitro diagnostic tests (IVDs), specifically rapid diagnostic tests (RDTs) for infectious diseases (ID), is essential for demonstrating their value and promoting their adoption in evidence-based practice. Unlike pharmaceuticals, the effectiveness of IVDs depends on how results are used to inform clinical decisions highlighting the need for diagnostic stewardship and implementation considerations throughout a test's lifecycle. In simple terms, traditional study methodologies designed for pharmaceutical interventions to evaluate IVDs will not comprehensively demonstrate their value [1].

The foundation for choosing an IVD depends primarily on the test's specifications, performance, and diagnostic accuracy [2]. However, unlike pharmaceuticals, which receive regulatory clearance based on their impact on clinical outcomes, IVDs are often approved and scaled up for public health use based on accuracy alone. From both a philosophical and practical standpoint, IVDs cannot be fully evaluated until their impact on patient outcomes is understood [3]. Many peripheral components of healthcare delivery affect this understanding, including how the test is implemented (ie, where, how, and with whom it is used), how it is interpreted, and the associated provider behaviors resulting from the test results. To ensure that findings are reliable, relevant, and generalizable, it is crucial to account for as many of these considerations as possible when studying the clinical impact of IVDs. Methodological best practices should be employed to address the complexities associated with measuring the impact of IVDs. Integrating principles of diagnostic excellence (including diagnostic stewardship) and antimicrobial stewardship (AMS) is important, as these principles have been shown to significantly impact care delivery and improve patient outcomes [4, 5].

Whenever possible, implementation science should be leveraged to bridge the gap between evidence and practice, adapting study designs to meet contextual needs and systematically addressing behavior change alongside IVD implementation [6]. Finally, efforts should be made to design studies that address the needs of populations often excluded from primary clinical trials due to methodological limitations and generate evidence that supports test reimbursement, policy changes, and integration into clinical guidelines. This review outlines key methodological considerations for studying the impact of these tests, the importance of diagnostic stewardship and implementation science in clinical research, and the data requirements needed for guideline development, test reimbursement, and enhanced market access.

METHODOLOGICAL CONSIDERATIONS FOR STUDYING AND DEMONSTRATING THE IMPACT OF DIAGNOSTIC TESTS

Studying the impact of IVDs requires careful consideration. Before beginning a study, it is essential to determine whether a clinically relevant opportunity for improvement exists for the primary outcome that the study is designed to assess (Figure 1). This should be established with local baseline quantitative data beyond a subjectively perceived opportunity to improve care (Table 1). Failing to perform such due diligence may lead to undetectable changes due to small baseline event rates and sample size limitations. Moreover, to ensure the capacity of a test for impact, and selection of relevant outcomes, study outcome selection should be driven by the use of existing or explicit development of potential care pathways, including considerations such as resource utilization and patient reported outcomes (eg, values and preferences) [7–9]. Guidance on developing such pathways may be found elsewhere [10]. To ensure clear and specific estimates of effect are established, a study should characterize a specified population, intervention, comparison, outcome, the timing in the care pathway of testing, and the clinical setting (PICOTS) [11]. Finally, in selecting study outcomes, subjective adjudicated primary outcomes or part of a primary composite outcome should generally be avoided given the often very low interrater and intrarater reliability of agreement in such approaches which usually biases results toward the null [12–14].

Figure 1.

Figure 1.

Example of a hypothetical directional acyclic graph (DAG) for in vitro diagnostic tests (IVDs) in bloodstream infections.

Table 1.

Methodological Considerations for IVD Clinical Outcomes Study Design

Step Key Points Example
Identify clinically relevant opportunities for improvement
  • Leverage baseline data to understand population and trends

  • Avoid subjective expert adjudicated primary outcomes where interrater reliability is low

  • Use PICOTS framework

A study assessing the impact of a diagnostic intervention on antibiotic avoidance versus SOC fails to show an impact due to high rate of baseline antibiotic avoidance (ie, in the setting of highly effective antimicrobial stewardship)
Leverage diagnostic stewardship principles in study design
  • Use knowledge of pretest probability and patient population to determine composition of intervention group

A diagnostic intervention is only randomized to patients in which the clinician indicates the test may change management
Consider heterogeneity and local context on test impact
  • Evaluate center-level effects, consider mixed-methods approaches

RCT ADEQUATE trial: RVP versus SOC demonstrated a 10% decrease in antibiotic prescribing overall, but when stratified by center, significant heterogeneity in effects seen
Utilize observational study design methods and the Target Trial Framework
  • Design the observational study to mimic an RCT

  • Help avoid common pitfalls with many biases (eg, immortal time bias)

RCT-DUPLICATE study: Demonstrated that this approach yields the same statistical significance in 75% of studies and the same estimate in 66% of studies
Include crude and adjusted estimates
  • Rather than using statistical methods for causal model selection, use approaches such as directional acyclic graphs (DAGs) and clinical judgement

Stepwise selection and overfitted models may lead to spurious associations and biased regression estimates
Use alternative approaches to analysis
  • Do not assess an outcome while conditioning on a future event

  • Address immortal time bias

  • Use intention-to-treat or per-protocol analysis, as appropriate

Conditioning on future events: Evaluating the efficacy of a diagnostic among only influenza-positive patients.
Immortal time bias: selecting patients who received the result of their test while in the ED versus all patients in the SOC arm for an ED length of stay measurement.

Abbreviations: ED, emergency department; PICOTS, population, intervention, comparison, outcome, timing, and clinical setting; SOC, standard of care; RCT, randomized controlled trial; RVP, respiratory viral panel.

Randomized controlled trials (RCTs) evaluating diagnostic interventions are inherently more complex in design and interpretation than traditional therapeutic trials. Despite clear evidence—from clinical practice and the literature—that local factors such as prescribing norms, guidelines, epidemiology, and provider attitudes strongly influence a test's impact, RCT conclusions often overlook these nuances [15]. Instead, they are often viewed as definitive statements on the generalizable use of RDTs. This disparity underscores a fundamental misunderstanding of diagnostic versus therapeutic research, prompting leading diagnostic methodologists to describe diagnostic RCTs as a “sheep in a wolf's clothing” [16]. Crucially, an RCT can only suggest that an IVD can improve care; it cannot definitively show that it will or will not do so under every clinical practice. The recent multicenter ADEQUATE trial comparing respiratory viral pathogen (RVP) testing with standard of care (SOC) in emergency departments (EDs) exemplifies this principle: although antibiotic prescribing decreased overall by about 10% (P < .05), individual centers exhibited highly heterogeneous effects—ranging from substantial reductions to null effects and even increases, all under the same protocol and PICOTS framework (Table 1) [17]. This variation illustrates how essential it is to measure center-level factors influencing impact, yet doing so typically requires resource-intensive multicenter, mixed-methods designs. However, these approaches have been shown to be possible with IVDs in large multicenter observational studies [18]. Research networks could further facilitate future endeavors in this area with RCTs or observational designs. However, given that diagnostic RCTs often cost upwards of a million dollars for single-center studies and nearly half of them fail to complete enrollment, questions remain about the utility of such design and its associated financial expenditures in research [19]. Given these challenges, well-designed observational studies present an appealing, more cost-effective alternative for assessing real-world impact.

While more efficient in resource requirements, future observational IVD studies will require modern epidemiological principles. Current systematic reviews report that approaches to RDT study design usually have moderate to substantial potential for bias and, thus, represent a low quality of evidence [20–23]. To ameliorate these biases, the application of the “target trial framework,” or designing an observational study to emulate RCT data measurement and analysis, facilitates the avoidance of common pitfalls with many biases (eg, immortal time bias) (Supplementary Figure 1). Moreover, as proof of concept, this approach has yielded the same statistical significance in 75% of RCTs and the same effect estimate in 66% of studies in the RCT-DUPLICATE Initiative of 32 RCTs—an initiative started under the 21st Century Cures Act to inform when observational studies could inform regulatory decisions [24].

Opportunities to enhance study design to reduce bias and improve the quality of evidence include providing not only basic estimates of effect but also confounding-adjusted estimates. Such adjusted estimates are typically missing in almost all observational studies of RDTs [20–23]. Moreover, when applied in the more general ID literature, adjustment models often neither specify their goal (ie, description, prediction, or causation), nor follow established recommendations for model development. This is important for the model design approach and study interpretation, so much so that the International Society for Pharmacoepidemiology (ISPE), the Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) Statement, and leading methodological experts in the field, all recommend not using statistical methods (eg, stepwise selection for a prediction model) for causal model covariate selection but instead using approaches like directional acyclic graphs (DAGs) for delineating causal relationships (Figure 2) and clinical judgment to determine causal model covariates for adjustment [25–29].

Figure 2.

Figure 2.

Decision schematic for designing IVD studies. Abbreviations: PICOTS, population, intervention, comparison, outcome, timeframe, setting; DAG, directed acyclic graph; IVD, in vitro diagnostic test; RCT, randomized controlled trial; MOST, multiphase optimization strategy trial; SMART, sequential multiple arm randomization trial; HARPER, HARmonized Protocol Template to Enhance Reproducibility.

Improving study design quality leads to decreasing potential bias and increased quality of evidence for informing clinical practice and guidelines. Additionally, these thoughtful approaches to embracing causal frameworks helps avoid adjusting for noncausal variables, which are known to bias estimates. For instance, hemodialysis is usually a confounder—causally associated with both the exposure and outcome—for antibiotic studies but would be an exclusion criterion in a study of procalcitonin. Similarly, penicillin allergy is a confounder for antibiotic studies but not for IVD studies. These examples delineate where misspecifying appropriate confounders for adjustment can lead to biased estimates of effect. Finally, observational studies are still limited akin to RCTs in answering “can it work” for a setting at a point in time rather than “does it work” given the changes inherent with diagnostic efficacy due to center-level effects of local practices (eg, business hours only coverage of RDT), availability of new therapies to improve outcomes with RDT (eg, Paxlovid, ceftazidime–avibactam), learning of RDT over time (eg, clinician will often ignore results of a test when learning to trust the test), and many other factors [30–32].

Finally, for both RCTs and observational studies, there are opportunities for more informative estimates of effect through alternative approaches to analyses. Often in the literature, studies report for a subset of test result patients, such as influenza-positive patients [21]. While these data are informative in elucidating mechanisms of clinical benefit of the test, they are conditioning on future unknowable information (ie, the result of a test) and, thus, are both not an intervention that can be deployed and unable to be appropriately quantified for cost-effectiveness and implementation decision making. Moreover, conditioning on future events may lead to immortal time bias such as selecting patients who received the result of their test while in the ED versus all patients in the SOC arm for an ED length of stay measurement, again highlighting the importance of rigorous modern epidemiological principles and expertise in study design [33]. In contrast to Dr Boten's suggestion that the inclusion criteria require a healthcare provider to acknowledge the test could influence their decision making, studies may alternatively use intention-to-treat (ITT) (ie, the process of analyzing study subject data based on their initial exposure or treatment assignment regardless of not receiving the intervention) as the primary analysis and per-protocol (ie, analysis only of study subjects who adhered strictly to the protocol) as the secondary analysis. This approach captures both a test's general, pragmatic impact and its more explanatory effects when results are employed. Similarly, modified ITT may help to avoid diluting the impact of the intervention when measured among noneligible patients for an outcome (eg, excluding coagulase-negative Staphylococcus [CoNS]) contaminants when measuring the mortality impact of RDT in bloodstream infections). Such analysis approaches beyond ITT have been performed historically for procalcitonin trials and have an opportunity to yield explanatory results of the impact of a test under ideal conditions. Recently, intention-to-diagnose has been proposed as a principle that is analogous to ITT used in clinical trials, wherein all enrolled and tested patients remain included in statistical analysis to ensure a more generalizable and less biased evaluation of IVDs performance [34]. The intention-to-diagnose principle includes the consideration for results that either fall into a diagnostic gray zone or are invalid due to application issues or scientific limitations of the IVD.

NOVEL METHODS FOR STUDYING THE CLINICAL IMPACT OF DIAGNOSTIC TESTS

When a new IVD is approved, the available literature to help clinicians assess their clinical impact on the diagnosis and management of infections is largely limited to in vitro studies comparing the new test to a prespecified gold standard [35]. Considerations regarding clinical decision making, antimicrobial resistance (AMR), and antimicrobial and diagnostic stewardship, goals should be assessed, while IVDs are being developed [36–38]. Frequently, however, there is little beyond comparison of sensitivity and specificity of on-panel targets to help clinicians determine which test may best suit their health system or patient population.

The Antibacterial Resistance Leadership Group (ARLG) is a national, federally funded, initiative whose mission is to design and execute clinical trials that enhance the prevention, diagnosis, and treatment of antimicrobial-resistant infections [39–41]. As such, the ARLG has developed novel methodological and statistical approaches to assist researchers in advancing optimal treatment of antibiotic-resistant infections, methods for prevention of AMR, streamlining comparative diagnostic research processes, and providing clinically relevant endpoints for clinicians. This section will review the ARLG-developed Benefit–Risk Evaluation of Diagnostics: A Framework (BED-FRAME), Desirability of Outcome Ranking Management of Antimicrobial Therapy (DOOR-MAT), and Desirability of Outcome Ranking (DOOR), with a comparison in Table 2. Each framework is further described below with examples of applications to IVDs.

Table 2.

Comparison of ARLG-Developed Methodologies That Can Be Applied to Diagnostic Studies

BED-FRAME DOOR-MAT DOOR
Full name
  • Benefit–Risk Evaluation of Diagnostics: A Framework (BED-FRAME)

  • Desirability of Outcome Ranking Management of Antimicrobial Therapy (DOOR-MAT)

  • Desirability of Outcome Ranking (DOOR)

Primary objective
  • To guide the assessment of the trade-offs between multiple IVDs and their potential impact on clinical decision making.

  • To optimize the management of antimicrobial therapy decisions based on organism resistance profiles.

  • To rank and evaluate the desirability of outcomes between treatment groups.

Methodology/data
  • Five-step approach.

  • Uses a benefit–risk framework to evaluate IVDs and inform clinical decisions.

  • Prevalence of organisms and AMR data necessary.

  • Relative clinical importance of missed identification/resistance needs to be determined through survey of clinicians.

  • Six step approach.

  • Extends the ordinal framework design to incorporate matrix-based management of antimicrobial therapy relative to phenotypic susceptibility profiles.

  • Uses data on antimicrobial therapy decisions, AMR profiles, and partial credit scoring analysis for quantitative comparisons.

  • Employs ordinal outcome ranking to determine the most effective therapy while balancing benefits and risks.

  • Pairwise comparisons indicate the likelihood that a randomly assigned patient in one group will experience better (or worse) outcomes than a patient in another group.

Application area
  • Direct evaluation of IVDs in the context of AMR.

  • Indirect evaluation of IVDs through ranking of antimicrobial therapy decisions, specifically with respect to phenotypic antimicrobial resistance.

  • Evaluation and ranking of clinical outcomes when comparing two different treatment strategies, ideally in a RCT.

Implementation
  • Requires a structured approach to assess diagnostics and their potential impact on therapy, includes direct comparisons, data abstraction, and survey development.

  • Focused on ranking and managing the desirability of outcomes in antimicrobial therapy. Includes indirect comparison through downstream decisions.

  • Focuses on determining the desirability of overall patient outcomes between two interventions in RCTs.

  • Includes indirect comparison through clinical outcomes.

Strengths
  • Helps clinicians evaluate whether IVDs are beneficial in the context of treatment options specific to their patient populations while incorporating diagnostic stewardship principles.

  • More granular assessment of complex therapy decisions, allowing for adjustments based on institutional resistance and preferences in antimicrobial therapies (incorporates AMS principles).

  • Uses real-world clinical outcomes data. Simplified ordinal approach allows for patient-based outcomes analysis.

  • Can still analyze individual outcome components.

Weaknesses
  • Does not directly address therapy decisions or clinical outcomes.

  • Challenging to determine relative clinical importance.

  • Statistical comparisons may be more challenging for clinicians to interpret.

  • Can oversimplify antimicrobial therapy decisions.

  • No standardized approach to spectrum of activity determination or partial credit scoring.

  • May be logistically difficult to compare diagnostics outside theoretical comparisons.

  • Few disease states have standardized DOORs.

  • The heterogeneity in what is included in each rank level can decrease generalizability

  • Identifying and defining clear, mutually exclusive outcomes can be difficult.

  • DOOR may oversimplify the decision-making process by reducing outcomes to a simple ranking, particularly with different types of outcomes.

  • Application in non-RCTs (retrospective cohort studies) is common but can introduce bias.

Abbreviations: AMR, antimicrobial resistance; AMS, antimicrobial stewardship; ARLG, Antibacterial Resistance Leadership Group; IVD, in vitro diagnostic test; RCT, randomized controlled trial.

Comparing In Vitro Diagnostics Use the Benefit–Risk Evaluation of Diagnostics: A Framework

The ARLG BED-FRAME was developed to assist in choosing between IVDs while considering the impact of local AMR prevalence and potential downstream clinical impact [42]. This framework was developed to be supplementary to traditional IVD evaluations. The BED-FRAME consists of five steps focusing on visualization of the test risks and benefits as localized to a specific population of interest. The ARLG provides examples of the application of components of BED-FRAME to compare IVDs that detect carbapenem-resistant Enterobacterales, Acinetobacter baumannii, and P. aeruginosa through the PRIMER studies [43–45]. Pragmatic analysis using the intention-to-diagnose principle is exemplified with BED-FRAME [34]. In particular, BED-FRAME could assist diagnostic stewardship and AMS programs in determining which IVDs would be best for their institutions prior to onboarding by considering patterns of resistance and how missed detection of that resistance could impact optimal patient management.

Published real-world examples of BED-FRAME outside the ARLG are limited. Smith et al [46] used a modified version of BED-FRAME to compare three RDT platforms for blood culture organism identification and resistance detection. This single-center study collected 301 monomicrobial-positive blood culture samples from unique patients. Pathogen prevalence was determined through local clinical microbiology data, and relative clinical importance and preferred diagnostic attributes were determined by surveying clinicians and AMS team members. The sensitivity of all three RDTs compared with culture results was over 88%. As determined by the AMS team, the most important diagnostic attributes were accurate pathogen identification (77.8%), identifying resistance determinants (16.7%), and time-to-results (5.6%). Following the steps of BED-FRAME, health systems can compare, through direct or theoretical application, different IVDs and balance stewardship priorities with laboratory workflow.

Assessing Antibiotic Use Through the Desirability of Outcome Ranking Management of Antimicrobial Therapy Framework

Many scoring tools have been developed to quantify antimicrobial spectrum of activity for intra- and interfacility benchmarking beyond traditional antimicrobial days of therapy [47–49]. Optimal or desirable therapy is generally considered the narrowest spectrum with in vitro activity, but this simplified definition is often difficult to apply pragmatically [50]. In particular, these scoring tools do not allow for consideration of AMR profiles and solely focus on spectrum [40, 51, 52]. The DOOR-MAT is a novel framework to assess antimicrobial therapy decisions considering both antimicrobial spectrum of activity and the organism AMR patterns. The DOOR-MAT cross-references the prescribed antimicrobial to the final organism resistance profile and applies a score of 0 (inactive) to 100 (active and most narrow). It is unique in its structure and ability to aid clinicians in comparing therapy decisions in a quantifiable manner given patient-specific characteristics and diagnostic or AMS goals. Studies have published examples of DOOR-MAT frameworks that can be applied for comparisons specific to Enterobacterales [51–53]. These can be used as a starting point and tailored to individual health system formularies or prescribing patterns.

Two studies have leveraged DOOR-MAT to compare theoretical antimicrobial therapy decisions based on different blood RDT panels [54, 55]. These studies adapted the DOOR-MAT developed by Wilson et al to a series of clinical blood culture isolates tested with different RDT panels. The results were compared to traditional culture and automated antimicrobial susceptibility testing. For each patient, RDT results were presented with case vignettes addressing clinical considerations such as age, infection severity, infection source, and history of AMR. Based on the results from the different panels, two clinicians made theoretical antimicrobial therapy decisions. With a high rate of agreement between prescribing clinicians, these decisions were then referenced against the DOOR-MAT framework considering the final phenotypic resistance profiles of the detected organisms. The studies highlighted how traditional metrics such as positive percent agreement (PPA) do not accurately reflect the impact of different panels on antimicrobial therapy decisions. For instance, while two panels showed similar PPAs {98% (95% confidence interval [CI] 94.3% to 99.6%) vs 98.6% (95% CI 95.1% to 99.8%)}, their mean DOOR-MAT scores differed significantly {a score of 91.9 (standard deviation [SD] ± 23.1) vs 86.8 (SD ± 28.5)} [52]. A second study compared two panels, one of which had expanded organism and resistance detection [51], and found that even though both had high sensitivity, the expanded panel had a much higher DOOR-MAT score (59.9 [SD ± 33.7] vs 89.7 [SD ± 24.7]), underscoring its role in in optimizing antimicrobial therapy.

Determining Patient Clinical Outcomes With the Desirability of Outcomes Ranking Framework

Both BED-FRAME and DOOR-MAT allow for direct or indirect comparison of IVDs, respectively, by considering local AMR profiles and incorporating diagnostic stewardship and AMS principles. They do not, however, allow for assessment of patient clinical outcomes. This is where the DOOR is useful [56]. The DOOR allows for holistic patient-centric comparisons of clinical outcomes by combining multiple endpoints (ie, efficacy and adverse events) in a ranked ordinal fashion. Patient outcomes are mutually ranked from range the best outcome (alive with no adverse events) to the worst outcome (mortality). After each patient is assigned a DOOR level, distribution of levels between groups is compared as is the overall probability of a better outcome.

The DOOR is often combined with a continuous variable to assist as a tiebreaker among the pairwise comparisons of DOOR levels. The most common tiebreaker is Response Adjusted for Duration of Antibiotic Risk, which evaluates antibiotic duration under the AMS principle that shorter is better [57]. When patients are assigned the same DOOR level, shorter duration is used to determine which patient has an overall better outcome. Other continuous variables that could potentially be leveraged include health-related quality-of-life measures or DOOR-MAT scoring to consider the spectrum of antimicrobial activity [53, 58].

The DOOR framework has mainly been applied to post hoc evaluations of RCTs comparing novel antimicrobials to SOC for infections such complicated intraabdominal infections or hospital-acquired pneumonia, multidrug-resistant infections (CRACKLE-2), as well as for comparing antimicrobial therapy regimens such as combination beta-lactam/vancomycin (CAMERA-2) and vancomycin area-under-the-curve (PROVIDE) [59–65]. One small quasiexperimental study used DOOR to compare patient outcomes pre- versus postimplementation of blood culture RDTs [66]. Researchers used a 3-level DOOR that was based on the ability to appropriately escalate or deescalate therapy according to the spectrum-of-activity scoring system. The authors reported a 58% (95% CI 54%–62%) overall probability of better outcomes in the post-RDT group compared with the pre-RDT group. Another small study compared the role beta-D glucan testing in optimizing therapy for candidemia in critically ill patients [67]. Numerous opportunities exist for large-scale high-quality IVD studies that incorporate DOOR as a primary or secondary endpoint.

The DOOR methodology has limitations, however, primarily due to its heterogeneous development and application [68]. This includes variation in the number of levels and endpoints considered within each level, as well as the study designs using DOOR. Some studies have developed DOORs with limited rigor, while others rely on consensus methods, such as modified Delphi techniques [66, 69]. Instead of developing DOORs for each study ad hoc, the ARLG has consensus-based DOORs for specific infections, such as Staphylococcus aureus BSI and HAP/VAP, which can be modified to enhance study validity [62, 70]. An additional potential limitation to the DOOR, however, is that it has the potential to obscure clinical outcomes by including multiple outcomes of different severity into a single level. Additionally, DOOR was originally developed and applied to results from RCTs, decreasing risk of bias in treatment or group assignments. Many published DOOR studies, however, have applied the method to retrospective cohort data, with or without additional statistical methods such as inverse probability of treatment weighting to help reduce risk of selection bias. Overall, more work is needed to standardize the methodological approach to allow investigators to develop disease-specific scores beyond what has been studied in a manner consistent with ARLG's previous tools.

Introduction to Hybrid Effectiveness–Implementation Study Framework

Hybrid effectiveness–implementation studies can improve timely use of novel IVDs while also considering key diagnostic stewardship metrics and outcomes. Proposed by Curran et al [71] in 2012, these study designs aim to bridge the gap between evidence-based interventions and practical application. Hybrid studies simultaneously assess the intervention effectiveness and implementation strategy, allowing researchers to evaluate not only how well an intervention works in real-world settings but also how it is integrated and adopted by the target population [71, 72]. There are three different types of hybrid designs, each balancing effectiveness and implementation with differing degrees of emphasis (Table 3). Initially intended for clinical patient-level RCTs, their scope has expanded to include cluster-randomized, stepped wedge and other quasiexperimental studies [73]. Curran et al offer guidance on selecting the right hybrid design that fits with specific types of research questions, considering factors such as existing evidence for the intervention, the extent of adoption necessary, and known barriers and facilitators to implementation. Regardless of the hybrid type and study design selected, leveraging hybrid effectiveness–implementation studies aligns perfectly with the core principles of diagnostic stewardship, which aims to optimize real-world application of diagnostic tools while ensuring they improve patient care [37].

Table 3.

Comparison and Diagnostic Test Examples of Hybrid Effectiveness–Implementation Studies

Hybrid Type Primary Focus Secondary Focus Example
Type 1 Effectiveness of an intervention Preliminary implementation data A patient-level RCT testing a novel blood RDT with phenotypic AST for FDA approval while documenting provider adoption barriers.
Type 2 Both effectiveness and implementation strategies N/A Cluster-randomized study of an available RDT for BSI in acute care settings while comparing different AMS strategies to adoption (eg, implementing a new RDT–treatment algorithm, educational tools).
Type 3 Implementation strategy Effectiveness of an intervention Comparing approved IVDs with paired strategies (eg, electronic health record [EHR]-based interventions) to reduce unnecessary C. difficile testing while assessing the impact on patient management.

Abbreviations: AST, automated susceptibility testing; FDA, Food and Drug Administration; RCT, randomized controlled trial; RDT, rapid diagnostic test.

Since their inception, hundreds of hybrid effectiveness–implementation studies of healthcare interventions have been conducted [73]. These interventions include expanding age-friendly healthcare in VA medical centers, implementing cognitive behavioral therapy, evaluating chronic pain, and more [74–77]. Few have focused on ID, and even fewer on implementing novel IVDs [78–80]. Table 3 provides brief examples of potential applications of hybrid study designs for IVDs. Another example could be a hybrid type 2 trial, using a quasiexperimental design assessing the impact and implementation of a point-of-care (POC) respiratory RDT in children. The coprimary outcomes would be the effect of the POC RDT on effectiveness (ie, decreased antimicrobial use), implementation (ie, acceptability, adoption, feasibility), service (ie, efficiency, equity, timeliness), and client (ie, satisfaction) outcomes. By collecting effectiveness metrics and implementation data concurrently, the hybrid design maximizes research efficiency, as they reduce the need for separate studies focused solely on either aspect. The conventional sequential research approach—from efficacy to effectiveness to implementation—can be time-consuming, delaying the practical application of an intervention. This integrated approach provides a more holistic understanding of an intervention's impact, offering insights into both its clinical outcomes and the practical challenges of its adoption, scalability, and sustainability. This is particularly valuable for interventions where the success of an intervention is often influenced by factors such as provider behavior and organizational context, which is often the case with novel diagnostic tests.

BROADER IMPACT OF EVIDENCE GENERATION FOR DIAGNOSTICS

The Fryback and Thornbury (FT) hierarchical framework of efficacy was originally created to appraise diagnostic imaging literature but has since been applied to other IVDs [81]. It consists of six levels, focusing on different aspects of diagnostic efficiency. Levels 1 and 2 focus on technical efficacy and diagnostic accuracy, metrics commonly evaluated during clinical trials and the FDA clearance process. Levels 3 through 6 encompass diagnostic thinking efficacy, therapeutic efficacy, patient outcome efficacy, and societal efficacy, respectively. Fryback and Thornbury emphasize that while each level is necessary, none is sufficient alone. For example, level 3 examines whether diagnostic information influences provider thinking, impacting level 4 (changes to patient management). This connection also affects levels 5 and 6, which evaluate patient outcomes and broader societal implications, respectively.

Major payers, such as the Centers for Medicare and Medicaid Services, utilize this framework, among others, to determine what is necessary and reasonable for reimbursement [82]. In vitro diagnostic test clinical trials (ie, premarket launch) often produce substantial data for FDA submissions that focus on technical and diagnostic efficacy (ie, levels 1 and 2 of the FT framework), which is crucial for safety and efficacy demonstration. However, these data may not suffice for payers’ coverage decisions. A review of Health Technology Assessment (HTA) methodologies demonstrated that while approaches may vary, there is a consensus that accuracy alone cannot establish clinical effectiveness [83]. The way evidence is generated is crucial for test adoption and reimbursement; a PICOTS framework, highlighted earlier in this paper, should always be utilized, and the role of a new IVD in the current care pathway should be demonstrated whenever possible.

Direct evidence, which involves comparing clinical outcomes following different test strategies within the same primary study, typically in the form of RCTs that randomize by test method, is preferred by payers and HTA agencies [83]. This is especially true for novel molecular IVDs, where providers often depend on indirect evidence, while payers rely on clinical guidelines and results from RCTs [84]. However, direct diagnostic evidence that includes treatment pathways (eg, test-plus-treatment RCTs) is rare and has considerable limitations. Reviews of test–treatment RCTs indicate that they are more susceptible to methodological challenges than RCTs for drugs [83]. Furthermore, the controlled nature and enrollment requirements of such trials can make them infeasible [85]. Conducting high-quality RCTs for IVDs is challenging due to factors like blinding, learning effects (eg, the user learning curve), selecting appropriate comparators, and changes in IVD performance over time.

Real-world Data and Real-world Evidence

To supplement RCT data and support clinical guideline development and payer coverage, real-world data (RWD) has emerged as a viable option to generate real-world evidence (RWE) [85, 86]. The 2024–2025 International Society for Pharmacoeconomics and Outcomes Research report lists RWE as the top trend in health research. Real-world evidence, which uses data from routine healthcare delivery (eg, electronic health records, insurance claims databases, data from wearable devices), offers a more realistic view of diagnostic interventions and outcomes and has gained support from various agencies, including the FDA, European Medicines Agency, National Institute for Health and Care Excellence, and Canada's Drug and Health Technology Agency. A 2018 Institute for Clinical and Economic Review (ICER) publication on RWE for coverage decisions highlights that the healthcare community often operates under outdated evidence hierarchies that regard RCTs as the single gold standard for evidence generation and healthcare decision making [87]. These hierarchies were not designed for a world driven by big data, where RWE can supplement RCTs and enhance our understanding of the effectiveness of interventions across various populations. While RWE is increasingly utilized in IVD regulatory decision making, hesitations remain in the postlaunch interventional and clinical outcomes space due to potential biases in nonrandomized studies and perceived flaws in study design [82, 88]. Ultimately, a lack of confidence in observational studies has slowed uptake of RWE into policy [89]. Therefore, studies must be rigorously designed to minimize bias. This rigor is crucial, as robust RWE plays a significant role in multiple healthcare decisions, including HTAs, payer decisions, clinical guidelines, and health policy [87, 90].

Real-world evidence may be generated using various study designs and with data acquired through routine health practice or real-world settings. The most common nonrandomized study types using RWE to assess comparative effects of IVDs and pharmaceuticals are observational cohort studies and single-arm trials with a real-world external control. These methodologies are often employed when full RCTs are uncommon and help address qualitative considerations, such as patient experiences with an IVD or intervention [91]. Additionally, pragmatic clinical trials, or trials that are randomized and contain varying levels of RWD elements, are recommended as strong alternative approaches [89]. Real-world evidence is used throughout an IVD product life cycle, including leveraging data to identify future innovation, clinical trials for FDA submission, effectiveness studies, and label expansion [92]. The FDA provides a summary document detailing 90 examples of when RWD was used for regulatory decision making in the premarket and postmarket processes for IVDs, including label expansions, surveillance, and validation [88].

The potential for using RWD to generate RWE after the launch of an IVD is significant. This includes assessing test effects in routine settings, conducting head-to-head comparisons, evaluating populations often excluded from RCTs, and exploring intervention effect variability [91]. An example of the use of RWD to assess the impact of IVDs in the postlaunch period is seen in a 2023 study by Moon et al [93] which utilized hospital discharge data from the PINC AI Healthcare Database to understand the demographic and clinical characteristics, epidemiology, healthcare resource utilization, and cost among adult outpatients with acute gastroenteritis visiting US health systems. Furthermore, a recent publication by Fitzke et al [94] highlights study examples where RWD was used to answer study questions in ID, preventative medicine, and HTAs.

To effectively adopt IVDs, healthcare decision makers require evidence that is reliable, relevant, and robust, addressing clinical benefits and cost-effectiveness while providing recommendations for specific subpopulations [95]. Each study should include a clear research question and relevant patient outcomes that reflect how patients feel, function, or how long they live. The Core Outcome Measure in Effectiveness Trials database can be useful for identifying important outcomes [89, 96]. Additionally, researchers should prespecify study plans with tools such as the HARmonized Protocol Template to Enhance Reproducibility (HARPER) and are encouraged to register their protocols before starting the study [89, 97]. It is important to select appropriate data and conduct a systematic search early on to address key gaps. Strategies to reduce bias in sampling should include stratified or purposive sampling. The analytic design should align with the research question and data characteristics, including outcome distribution and sample size. During the study, diagnostic data checks should identify and mitigate potential biases, such as inconsistent variable definitions or missing data. Finally, transparency in reporting methods is crucial, especially when indirect or data-linking evidence methods are used. Involvement of key stakeholders, particularly payers, throughout the study design process is essential for gaining support and access approval [82, 84, 85, 98]. Finally, utilizing methods that support reliable causal inference alongside RWE is vital for building the “totality of evidence” for the IVD [99].

Where to Start When Designing Clinical Outcomes Studies for In Vitro Diagnostic Test

Designing a clinical outcomes study for IVDs can be challenging for general ID clinicians or clinical microbiologists. While engaging in clinical trial design and conducting performance studies are essential for the clearance and use of IVD products, this section will focus on designing clinical impact and utility studies conducted after a product has been launched and is available on the market (Figure 2). The foundation of any study lies in its research question and objectives. A crucial first step is to understand baseline data, which include existing evidence, gaps, patient populations of interest, and relevant epidemiological factors. This information is vital for identifying PICOTS and creating a conceptual model or DAG that clarifies the pathways and processes influencing study design and outcomes.

It is essential to involve a biostatistician as early as possible, especially when defining the study's outcomes, determining the appropriate study power, and calculating the necessary sample size to accurately measure those outcomes. The researcher should consider whether a single outcome or a composite outcome would more appropriately address the research question. In cases of AMR-focused research, employing composite outcome methodology, like the DOOR framework, may effectively convey a more comprehensive clinical narrative. Once the study's power, sample size, and outcomes are established, the researcher should also consider whether implementation science methodologies could enhance the study (Figure 2). Although implementation science is a relatively new field in clinical research, especially in the context of ID IVD studies, it offers valuable insights into how a test may be adopted and the contextual factors impacting clinical outcomes. Understanding that the clinical outcomes linked to IVD test results are influenced by human behavior is key; consequently, accounting for human behavior, psychology, and the local environment is crucial for interpreting these outcomes. A publication by Lane-Fall et al [100] encourages researchers to position themselves within the translational science pipeline while designing their studies, identifying moments when integrating implementation science methodologies might be beneficial.

Researchers should evaluate whether the effectiveness of the IVD has been established in the literature. For IVDs, this typically means demonstrating diagnostic accuracy and some degree of clinical impact. If there are significant data on performance and clinical utility (affirmative) or considerable data on accuracy and safety but incomplete data on clinical effectiveness (partially affirmative), the researcher should collect implementation-related data to improve adoption, access, or contextualize research findings. If there are no supportive data, indicating that the intervention is not yet ready for implementation, the researcher should refrain from incorporating implementation science methods and focus solely on demonstrating effectiveness. In instances where effectiveness is only partially demonstrated, hybrid implementation–effectiveness designs may be appropriate; guidance on designing such trials is available in other resources [71, 72].

Next, regardless of whether implementation science methodologies are used, researchers should assess whether a RCT is reasonable, ethical, and suitable. While RCTs are widely regarded as the gold standard in clinical research, certain scenarios may not favor them as the optimal design for studying the impact of an IVD. If an RCT is deemed the best design choice, researchers should adhere to best practices for RCTs and strive to enhance the generalizability and applicability of the study findings, including integrating implementation science methodologies. Furthermore, for complex interventions with multiple components, such as AMS interventions, researchers should seek to understand the impact of individual components rather than evaluating the intervention through post hoc analyses. In such cases, the Multiphase Optimization Strategy (MOST) can be used to analyze the effects of individual intervention components before conducting the RCT [101]. The MOST involves three phases (eg, screening, refining, and confirming) to identify components most likely to have a significant impact while meeting predefined criteria (eg, cost). Another innovative design, the Sequential Multiple Assignment Randomized Trial (SMART), may be beneficial for evaluating time-varying adaptive interventions, helping researchers understand the influence of the sequence of intervention components and the optimal timing for each component. Both the MOST and SMART are advanced design methods applicable to interventions that involve behavioral elements, as seen with IVDs. Detailed considerations and steps for implementing these methods have been published elsewhere [102].

If a researcher concludes that a RCT is not feasible, ethical, or suitable for addressing the study question, they should consider observational and quasiexperimental designs. It is important to recognize that data from observational RWE can effectively complement or expand on data from RCTs. Observational studies, especially retrospective ones that utilize large datasets (such as claims databases), can be particularly valuable for research focusing on health economics outcomes. Pragmatic trials, which are prospective randomized trials that leverage RWD, can help overcome some limitations of observational studies and improve the generalizability of findings [103]. However, challenges may arise related to feasibility and the balance between internal and external generalizability. Tools such as HARPER are available to guide researchers in designing RWE studies that enhance transparency and facilitate reproducibility. Furthermore, investigators may consider the use of approaches intended to help overcome significant limitations of observational research, such as the target trial emulation framework. One way to ensure that the analyses in observational studies preserve the optimal features of RCTs is to design them so that they emulate a hypothetical RCT [104]. To do this, the investigator should write the study protocol to include key elements such as eligibility criteria, treatment strategies, treatment assignment, the start and end of follow-up, outcomes, and the data analysis plan. Next, observational data can be used to identify eligible individuals and assign them to a treatment strategy according to their data and match the study analysis to what would be done during an RCT apart from adjusting for baseline confounders, which is done to emulate random treatment assignment [104, 105]. Taking this approach encourages investigators to align treatment assignment and follow-up at time zero, which helps avoid significant biases seen in observational research such as immortal time bias, lead time bias, and selection bias [105].

Several robust quasiexperimental designs exist that provide rigor while requiring fewer resources, making them excellent options for studies on clinical impact and implementation science of IVDs. The three most commonly utilized designs are interrupted time series, stepped wedge, and pre–post with a nonequivalent control. These designs are effective for assessing outcomes before and after the implementation of an IVD or related intervention and can incorporate methodologies from implementation science [106]. Advanced quasiexperimental approaches build on basic quasiexperimental methodologies by incorporating techniques that improve causal inference [107]. For example, a pre–post design with a control group and a difference-in-differences analysis addresses underlying time-dependent trends, whereas a simple pre–post design does not [108].

CONCLUSIONS

Evaluating the clinical impact of IVDs is complex and closely tied to various factors such as clinician behavior, treatment pathways, culture, and learning, requiring innovative approaches to study design and methodological rigor. It is essential to consider and adjust for the local context and to involve key stakeholders throughout all stages of study design and execution, including payors and policymakers when appropriate. Those engaged in clinical outcomes research related to IVDs should strive to advance the field by incorporating these methodologies and demonstrating their effectiveness. A thoughtful and comprehensive approach to studying the impact of IVDs has the potential to enhance patient care, improve public health, and increase access to better diagnostic testing.

Supplementary Material

ofaf489_Supplementary_Data

Notes

Financial support. Funding for this manuscript was provided by bioMérieux. K. C. and T. T. T. did not receive financial support for their contributions.

Supplement sponsorship. This article appears as part of the supplement “Advancing Infectious Diseases Diagnostics Through Research, Policy, and Market Access in the US,” sponsored by bioMérieux, Inc.

Contributor Information

Kimberly C Claeys, Department of Pharmacy Science and Health Outcomes Research, University of Maryland School of Pharmacy, Baltimore, Maryland, USA.

Andrea M Prinzi, bioMérieux, Medical Affairs, Salt Lake City, Utah, USA.

Tristan T Timbrook, Department of Pharmacy, Barnes-Jewish Hospital, Saint Louis, Missouri, USA.

Supplementary Data

Supplementary materials are available at Open Forum Infectious Diseases online. Consisting of data provided by the authors to benefit the reader, the posted materials are not copyedited and are the sole responsibility of the authors, so questions or comments should be addressed to the corresponding author.

References

  • 1. Baghdadi  JD, Morgan  DJ. Diagnostic tests should be assessed for clinical impact. CMI Commun  2024; 1:105010. [Google Scholar]
  • 2. Vetter  TR, Schober  P, Mascha  EJ. Diagnostic testing and decision-making: beauty is not just in the eye of the beholder. Anesth Analg  2018; 127:1085–91. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Sonke  GS, Verbeek  ALM, Kiemeney  LALM. A philosophical perspective supports the need for patient-outcome studies in diagnostic test evaluation. J Clin Epidemiol  2009; 62:58–61. [DOI] [PubMed] [Google Scholar]
  • 4. Core Elements of Hospital Diagnostic Excellence (DxEx) . 2025. Available at: https://www.cdc.gov/patient-safety/hcp/hospital-dx-excellence/index.html. Accessed 4 March 2025.
  • 5. CDC . Core Elements of Hospital Antibiotic Stewardship Programs | Antibiotic Use. 2021. Available at: https://www.cdc.gov/antibiotic-use/hcp/core-elements/hospital.html. Accessed 21 February 2022.
  • 6. Burnham  JP, Geng  E, Venkatram  C, Colditz  GA, McKay  VR. Putting the dissemination and implementation in infectious diseases. Clin Infect Dis  2020; 71:218–25. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Mustafa  RA, Wiercioch  W, Cheung  A, et al.  Decision making about healthcare-related tests and diagnostic test strategies. Paper 2: a review of methodological and practical challenges. J Clin Epidemiol  2017; 92:18–28. [DOI] [PubMed] [Google Scholar]
  • 8. Mustafa  RA, Wiercioch  W, Ventresca  M, et al.  Decision making about healthcare-related tests and diagnostic test strategies. Paper 5: a qualitative study with experts suggests that test accuracy data alone is rarely sufficient for decision making. J Clin Epidemiol  2017; 92:47–57. [DOI] [PubMed] [Google Scholar]
  • 9. Thompson  MJ, Suchsland  MZ, Hardy  V, et al.  Patient-centred outcomes of imaging tests: recommendations for patients, clinicians and researchers. BMJ Qual Saf  2023; 32:536–45. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Tuut  MK, Gopalakrishna  G, Leeflang  MM, et al.  Co-creation of a step-by-step guide for specifying the test-management pathway to formulate focused guideline questions about healthcare related tests. BMC Med Res Methodol  2024; 24:241. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Riva  JJ, Malik  KMP, Burnie  SJ, Endicott  AR, Busse  JW. What is your research question? An introduction to the PICOT format for clinicians. J Can Chiropr Assoc  2012; 56:167–71. [PMC free article] [PubMed] [Google Scholar]
  • 12. Schwartz  DN, Wu  US, Lyles  RD, et al.  Lost in translation? Reliability of assessing inpatient antimicrobial appropriateness with use of computerized case vignettes. Infect Control Hosp Epidemiol  2009; 30:163–71. [DOI] [PubMed] [Google Scholar]
  • 13. Pickens  CI, Gao  CA, Bodner  J, et al.  An adjudication protocol for severe pneumonia. Open Forum Infect Dis  2023; 10:ofad336. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Stern  SE, Christensen  MA, Nevers  MR, et al.  Electronic surveillance criteria for non-ventilator-associated hospital-acquired pneumonia: assessment of reliability and validity. Infect Control Hosp Epidemiol  2023; 44(11):1769–75. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Loubet  P, Fourati  S, Bouzid  D. Should multiplex PCR testing be integrated into antimicrobial stewardship programs for paediatric community-acquired pneumonia in the emergency department?  Clin Microbiol Infect  2025; 31:5–7. [DOI] [PubMed] [Google Scholar]
  • 16. Ferrante di Ruffano  L, Deeks  JJ. Test-treatment RCTs are sheep in wolves’ clothing (letter commenting on: J Clin Epidemiol. 2014;67:612–21). J Clin Epidemiol  2016; 69:266–7. [DOI] [PubMed] [Google Scholar]
  • 17. Bielicki  J. The value of in vitro diagnostic tests to reduce antimicrobial resistance and to improve patient outcomes. 2024. Available at: https://channel.ersnet.org/\\channel.ersnet.org/container-detail.php?id=108528. Accessed 27 March 2025.
  • 18. Britt  NS, Khader  K, He  T, et al.  Examining the clinical impact of rapid multiplex polymerase chain reaction-based diagnostic testing for bloodstream infections in a national cohort of the Veterans Health Administration. Pharmacotherapy  2022; 43(1):24–34. [DOI] [PubMed] [Google Scholar]
  • 19. Walters  SJ, Bonacho Dos Anjos Henriques-Cadby  I, Bortolami  O, et al.  Recruitment and retention of participants in randomised controlled trials: a review of trials funded and published by the United Kingdom Health Technology Assessment Programme. BMJ Open  2017; 7:e015276. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Peri  AM, Chatfield  MD, Ling  W, Furuya-Kanamori  L, Harris  PNA, Paterson  DL. Rapid diagnostic tests and antimicrobial stewardship programs for the management of bloodstream infection: what is their relative contribution to improving clinical outcomes? A systematic review and network meta-analysis. Clin Infect Dis  2024; 79(2):502–15. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Clark  TW, Lindsley  K, Wigmosta  TB, et al.  Rapid multiplex PCR for respiratory viruses reduces time to result and improves clinical care: results of a systematic review and meta-analysis. J Infect  2023; 86:462–75. [DOI] [PubMed] [Google Scholar]
  • 22. Timbrook  TT, Morton  JB, McConeghy  KW, Caffrey  AR, Mylonakis  E, LaPlante  KL. The effect of molecular rapid diagnostic testing on clinical outcomes in bloodstream infections: a systematic review and meta-analysis. Clin Infect Dis  2017; 64:15–23. [DOI] [PubMed] [Google Scholar]
  • 23. Hueth  KD, Thompson-Leduc  P, Totev  TI, et al.  Assessment of the impact of a meningitis/encephalitis panel on hospital length of stay: a systematic review and meta-analysis. Antibiotics (Basel)  2022; 11:1028. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Wang  SV, Schneeweiss  S, Franklin  JM, et al.  Emulation of randomized clinical trials with nonrandomized database analyses: results of 32 clinical trials. JAMA  2023; 329:1376–85. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Greenland  S, Neutra  R. Control of confounding in the assessment of medical technology. Int J Epidemiol  1980; 9:361–7. [DOI] [PubMed] [Google Scholar]
  • 26. Vandenbroucke  JP, von Elm  E, Altman  DG, et al.  Strengthening the Reporting of Observational Studies in Epidemiology (STROBE): explanation and elaboration. Epidemiology  2007; 18:805–35. [DOI] [PubMed] [Google Scholar]
  • 27. Robins  JM. Data, design, and background knowledge in etiologic inference. Epidemiology  2001; 12:313–20. [DOI] [PubMed] [Google Scholar]
  • 28. Varon  B, Palacios-Baena  ZR, de Kraker  MEA, et al.  Universal risk factors for mortality in bloodstream infections (UNIFORM): a systematic review and Delphi survey. Clin Microbiol Infect  2024; 30:453–61. [DOI] [PubMed] [Google Scholar]
  • 29. Public Policy Committee, International Society of Pharmacoepidemiology . Guidelines for good pharmacoepidemiology practice (GPP). Pharmacoepidemiol Drug Saf  2016; 25:2–10. [DOI] [PubMed] [Google Scholar]
  • 30. Gawrys  GW, Tun  K, Jackson  CB, et al.  The impact of rapid diagnostic testing, surveillance software, and clinical pharmacist staffing at a large community hospital in the management of Gram-negative bloodstream infections. Diagn Microbiol Infect Dis  2020; 98:115084. [DOI] [PubMed] [Google Scholar]
  • 31. Hueth  KD, Prinzi  AM, Timbrook  TT. Diagnostic stewardship as a team sport: interdisciplinary perspectives on improved implementation of interventions and effect measurement. Antibiotics (Basel)  2022; 11:250. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Satlin  MJ, Chen  L, Gomez-Simmonds  A, et al.  Impact of a rapid molecular test for Klebsiella pneumoniae carbapenemase and ceftazidime-avibactam use on outcomes after bacteremia caused by carbapenem-resistant enterobacterales. Clin Infect Dis  2022; 75:2066–75. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Rao  S, Lamb  MM, Moss  A, et al.  Effect of rapid respiratory virus testing on antibiotic prescribing among children presenting to the emergency department with acute respiratory illness: a randomized clinical trial. JAMA Netw Open  2021; 4:e2111836. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Evans  SR, Pennello  G, Zhang  S, et al.  Intention-to-diagnose and distinct research foci in diagnostic accuracy studies. Lancet Infect Dis  2025; 25(8):e472–81. [DOI] [PubMed] [Google Scholar]
  • 35. Caliendo  AM, Gilbert  DN, Ginocchio  CC, et al.  Better tests, better care: improved diagnostics for infectious diseases. Clin Infect Dis  2013; 57 Suppl 3:S139–70. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Zakhour  J, Haddad  SF, Kerbage  A, et al.  Diagnostic stewardship in infectious diseases: a continuum of antimicrobial stewardship in the fight against antimicrobial resistance. Int J Antimicrob Agents  2023; 62:106816. [DOI] [PubMed] [Google Scholar]
  • 37. Fabre  V, Davis  A, Diekema  DJ, et al.  Principles of diagnostic stewardship: a practical guide from the Society for Healthcare Epidemiology of America Diagnostic Stewardship Task Force. Infect Control Hosp Epidemiol  2023; 44:178–85. [DOI] [PubMed] [Google Scholar]
  • 38. Ku  TSN, Al Mohajer  M, Newton  JA, et al.  Improving antimicrobial use through better diagnosis: the relationship between diagnostic stewardship and antimicrobial stewardship. Infect Control Hosp Epidemiol  2023; 44:1901–8. [DOI] [PubMed] [Google Scholar]
  • 39. Chambers  HF, Evans  SR, Patel  R, et al.  Antibacterial resistance leadership group 2.0: back to business. Clin Infect Dis  2021; 73:730–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40. Evans  SR, Patel  R, Hamasaki  T, et al.  The future ain’t what it used to be…out with the old…in with the better: antibacterial resistance leadership group innovations. Clin Infect Dis  2023; 77:S321–30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Chambers  HF, Cross  HR, Souli  M, et al.  The antibacterial resistance leadership group: scientific advancements and future directions. Clin Infect Dis  2023; 77:S279–87. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Evans  SR, Pennello  G, Pantoja-Galicia  N, et al.  Benefit-risk evaluation for diagnostics: a framework (BED-FRAME). Clin Infect Dis  2016; 63:812–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Evans  SR, Hujer  AM, Jiang  H, et al.  Informing antibiotic treatment decisions: evaluating rapid molecular diagnostics to identify susceptibility and resistance to carbapenems against Acinetobacter spp. in PRIMERS III. J Clin Microbiol  2017; 55:134–44. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44. Evans  SR, Hujer  AM, Jiang  H, et al.  Rapid molecular diagnostics, antibiotic treatment decisions, and developing approaches to inform empiric therapy: PRIMERS I and II. Clin Infect Dis  2016; 62:181–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. Evans  SR, Tran  TTT, Hujer  AM, et al.  Rapid molecular diagnostics to inform empiric use of ceftazidime/avibactam and ceftolozane/tazobactam against Pseudomonas aeruginosa: PRIMERS IV. Clin Infect Dis  2019; 68:1823–30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Smith  RD, Zhan  M, Zhang  S, et al.  Comparison of three rapid diagnostic tests for bloodstream infections using Benefit-risk Evaluation Framework (BED-FRAME). J Clin Microbiol  2024; 62:e0109623. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Kakiuchi  S, Livorsi  DJ, Perencevich  EN, et al.  Days of antibiotic Spectrum coverage: a novel metric for inpatient antibiotic consumption. Clin Infect Dis  2022; 75:567–76. [DOI] [PubMed] [Google Scholar]
  • 48. Kanda  N, Ohbe  H, Jo  T, et al.  Trends in inpatient antimicrobial consumption using days of therapy and days of antibiotic spectrum coverage: a nationwide claims database study in Japan. J Infect Chemother  2024; 30:228–35. [DOI] [PubMed] [Google Scholar]
  • 49. Gerber  JS, Hersh  AL, Kronman  MP, Newland  JG, Ross  RK, Metjian  TA. Development and application of an antibiotic Spectrum Index for benchmarking antibiotic selection patterns across hospitals. Infect Control Hosp Epidemiol  2017; 38:993–7. [DOI] [PubMed] [Google Scholar]
  • 50. Ilges  D, Tande  AJ, Stevens  RW. A broad Spectrum of possibilities: spectrum scores as a unifying metric of antibiotic utilization. Clin Infect Dis  2023; 77:167–73. [DOI] [PubMed] [Google Scholar]
  • 51. Wilson  BM, Jiang  Y, Jump  RLP, et al.  Desirability of outcome ranking for the management of antimicrobial therapy (DOOR MAT): a framework for assessing antibiotic selection strategies in the presence of drug resistance. Clin Infect Dis  2021; 73:344–50. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Perez  F, Colindres  RV, Wilson  BM, et al.  Desirability of outcome ranking for the management of antimicrobial therapy (DOOR MAT) reveals improvements in the treatment of bloodstream infections caused by Escherichia coli and Klebsiella pneumoniae in patients from the Veterans Health Administration. Clin Infect Dis  2021; 73:1231–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Moon  RC, MacVane  SH, David  J, Morton  JB, Rosenthal  N, Claeys  KC. Clinical outcomes of early phenotype-desirable antimicrobial therapy for Enterobacterales bacteremia. JAMA Netw Open  2024; 7:e2451633. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Claeys  KC, Schlaffer  K, Smith  R, et al.  Day at the races: comparing BioFire FilmArray blood culture ID panels to verigene blood culture in gram-negative bloodstream infections using DOOR-MAT analysis. Clin Infect Dis  2021; 73(6):1103–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55. Claeys  KC, Hopkins  TL, Schlaffer  K, et al.  Comparing the clinical utility of rapid diagnostics for treatment of bloodstream infections using desirability of outcome ranking approach for the management of antibiotic therapy (DOOR-MAT). Antimicrob Agents Chemother  2021; 65:e0044121. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56. Evans  SR, Rubin  D, Follmann  D, et al.  Desirability of outcome ranking (DOOR) and response adjusted for duration of antibiotic risk (RADAR). Clin Infect Dis  2015; 61:800–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57. Schweitzer  VA, van Smeden  M, Postma  DF, Oosterheert  JJ, Bonten  MJM, van Werkhoven  CH. Response adjusted for days of antibiotic risk (RADAR): evaluation of a novel method to compare strategies to optimize antibiotic use. Clin Microbiol Infect  2017; 23:980–5. [DOI] [PubMed] [Google Scholar]
  • 58. Doernberg  SB; MAS for the Antibacterial Resistance Leadership Group . Desirability of outcome ranking and quality life measurement for antimicrobial research in organ transplantation. Transpl Infect Dis  2022; 24:e13888. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59. Kinamon  T, Gopinath  R, Waack  U, et al.  Exploration of a potential desirability of outcome ranking endpoint for complicated intra-abdominal infections using 9 registrational trials for antibacterial drugs. Clin Infect Dis  2023; 77:649–56. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60. Williams  DJ, Creech  CB, Walter  EB, et al.  Short- vs standard-course outpatient antibiotic therapy for community-acquired pneumonia in children: the SCOUT-CAP randomized clinical trial. JAMA Pediatr  2022; 176:253–61. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61. van Duin  D, Arias  CA, Komarow  L, et al.  Molecular and clinical epidemiology of carbapenem-resistant enterobacterales in the USA (CRACKLE-2): a prospective cohort study. Lancet Infect Dis  2020; 20:731–41. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62. Howard-Anderson  J, Hamasaki  T, Dai  W, et al.  Moving beyond mortality: development and application of a desirability of outcome ranking (DOOR) endpoint for hospital-acquired bacterial pneumonia and ventilator-associated bacterial pneumonia. Clin Infect Dis  2024; 78:259–68. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63. Petersiel  N, Davis  JS, Meagher  N, et al.  Combination of antistaphylococcal β-lactam with standard therapy compared to standard therapy alone for the treatment of methicillin-resistant Staphylococcus aureus bacteremia: a post hoc analysis of the CAMERA2 trial using a desirability of outcome ranking approach. Open Forum Infect Dis  2024; 11:ofae181. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64. Lodise  TP, Rosenkranz  SL, Finnemeyer  M, et al.  The Emperor's new clothes: prospective observational evaluation of the association between initial vancomycin exposure and failure rates among adult hospitalized patients with MRSA bloodstream infections (PROVIDE). Clin Infect Dis  2019; 70:1536–45. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65. Howard-Anderson  J, Hamasaki  T, Dai  W, et al.  Improving traditional registrational trial End points: development and application of a desirability of outcome ranking End point for complicated urinary tract infection clinical trials. Clin Infect Dis  2023; 76:e1157–65. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66. Smoke  SM, Patel  VV, Leonida  NI. The DOOR to antibiotic stewardship: refining assessments of interventions with desirability of outcome ranking. J Pharm Pract  2022; 35:403–6. [DOI] [PubMed] [Google Scholar]
  • 67. Giacobbe  DR, Signori  A. Interpreting desirability of outcome ranking (DOOR) analyses in observational studies in infectious diseases: caution still needed. Eur J Clin Microbiol Infect Dis  2019; 38:1985–6. [DOI] [PubMed] [Google Scholar]
  • 68. Ong  SWX, Petersiel  N, Loewenthal  MR, Daneman  N, Tong  SYC, Davis  JS. Unlocking the DOOR-how to design, apply, analyse, and interpret desirability of outcome ranking endpoints in infectious diseases clinical trials. Clin Microbiol Infect 2023; 29:1024–30. [DOI] [PubMed] [Google Scholar]
  • 69. Johns  BP, Dewar  DC, Loewenthal  MR, et al.  A desirability of outcome ranking (DOOR) for periprosthetic joint infection—a Delphi analysis. J Bone Jt Infect  2022; 7:221–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70. Doernberg  SB, Tran  TTT, Tong  SYC, et al.  Good studies evaluate the disease while great studies evaluate the patient: development and application of a desirability of outcome ranking endpoint for Staphylococcus aureus bloodstream infection. Clin Infect Dis  2019; 68:1691–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71. Curran  GM, Bauer  M, Mittman  B, Pyne  JM, Stetler  C. Effectiveness-implementation hybrid designs. Med Care  2012; 50:217–26. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72. Landes  SJ, McBain  SA, Curran  GM. An introduction to effectiveness-implementation hybrid designs. Psychiatry Res  2019; 280:112513. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73. Curran  GM, Landes  SJ, McBain  SA, et al.  Reflections on 10 years of effectiveness-implementation hybrid studies. Front Health Serv  2022; 2:1053496. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74. Salloum  RG, Bilello  L, Bian  J, et al.  Study protocol for a type III hybrid effectiveness-implementation trial to evaluate scaling interoperable clinical decision support for patient-centered chronic pain management in primary care. Implement Sci  2022; 17:44. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75. Siddiqi  A, Monton  O, Woods  A, et al.  Dissemination and implementation of a community health worker intervention for disparities in palliative care (DeCIDE PC): a study protocol for a hybrid type 1 randomized controlled trial. BMC Palliat Care  2023; 22:139. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76. Piazza  KM, Ashcraft  LE, Rose  L, et al.  Study protocol: type III hybrid effectiveness-implementation study implementing age-friendly evidence-based practices in the VA to improve outcomes in older adults. Implement Sci Commun  2023; 4:57. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77. Yanez  B, Czech  KA, Buitrago  D, et al.  Effectiveness and implementation of an electronic health record-integrated digital health intervention for managing depressive symptoms in ambulatory oncology: the My Well-Being Guide study rationale and protocol. Contemp Clin Trials  2023; 127:107121. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78. Wolfensberger  A, Clack  L, von Felten  S, et al.  Prevention of non-ventilator-associated hospital-acquired pneumonia in Switzerland: a type 2 hybrid effectiveness-implementation trial. Lancet Infect Dis  2023; 23:836–46. [DOI] [PubMed] [Google Scholar]
  • 79. Lam  RPK, Hung  KKC, Lui  CT, et al.  Early sepsis care with the National Early Warning Score 2-guided Sepsis Hour-1 Bundle in the emergency department: hybrid type 1 effectiveness-implementation pilot stepped wedge randomised controlled trial (NEWS-1 TRIPS) protocol. BMJ Open  2024; 14:e080676. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80. Van Pelt  AE, Bilker  WB, Nkwihorez  H, et al.  Increasing antiretroviral therapy adherence and retention in care among adults living with HIV in Philadelphia: a study protocol for a stepped-wedge cluster-randomised type 2 hybrid effectiveness-implementation trial of managed problem-solving plus (MAPS+) delivered by community health workers. BMJ Open  2023; 13:e079585. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81. Fryback  DG, Thornbury  JR. The efficacy of diagnostic imaging. Med Decis Mak  1991; 11:88–94. [DOI] [PubMed] [Google Scholar]
  • 82. Food and Drug Administration, Center for Devices and Radiological Health. Medical Device Coverage Initiatives: Connecting with Payors via the Payor Communication Task Force . FDA 2025; Available at: https://www.fda.gov/about-fda/cdrh-innovation/medical-device-coverage-initiatives-connecting-payors-payor-communication-task-force. Accessed 19 March 2025.
  • 83. Ferrante di Ruffano  L, Harris  IM, Zhelev  Z, et al.  Health technology assessment of diagnostic tests: a state of the art review of methods guidance from international organizations. Int J Technol Assess Health Care  2023; 39:e14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84. Deverka  PA, Douglas  MP, Phillips  KA. Use of real-world evidence in US payer coverage decision-making for next-generation sequencing-based tests: challenges, opportunities, and potential solutions. Value Health  2020; 23:540–50. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85. Dang  A. Real-world evidence: a primer. Pharm Med  2023; 37:25–36. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86. Top 10 HEOR Trends . Available at: https://www.ispor.org/heor-resources/top-10-heor-trends. Accessed 19 March 2025.
  • 87. Hampson  G, Towse  A, Dreitlein  WB, Henshall  C, Pearson  SD. Real-world evidence for coverage decisions: opportunities and challenges. J Comp Eff Res  2018; 7:1133–43. [DOI] [PubMed] [Google Scholar]
  • 88. Food and Drug Administration (FDA) . Examples of Real-World Evidence (RWE) Used in Medical Device Regulatory Decisions. Available at: https://www.fda.gov/media/146258/download. Accessed 10 March 2025.
  • 89. Berger  ML, Sox  H, Willke  RJ, et al.  Good practices for real-world data studies of treatment and/or comparative effectiveness: recommendations from the joint ISPOR-ISPE Special Task Force on real-world evidence in health care decision making. Pharmacoepidemiol Drug Saf  2017; 26:1033–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90. Gagne  J, Coplan  P, Davis  K, Qui  H. Advancing Patient Health Through Regulatory-Grade Real-World Evidence Generation 2024; Available at: https://www.ispor.org/publications/journals/value-outcomes-spotlight/vos-archives/issue/view/healthcare-systems-around-the-world/advancing-patient-health-through-regulatory-grade-real-world-evidence-generation. Accessed 1 March 2025.
  • 91. National Institute for Health and Care Excellence . Overview | NICE real-world evidence framework | Guidance | NICE. NICE 2022. Available at: https://www.nice.org.uk/corporate/ecd9/chapter/overview. Accessed 19 March 2025.
  • 92. Hennessy  S, Atsuta  Y, Hill  S, Rägo  L, Juhaeri  J. Council for International Organizations of Medical Sciences (CIOMS) Working Group XIII. Real-world data and real-world evidence in regulatory decision making: report summary from the council for international organizations of medical sciences (CIOMS) working group XIII. Pharmacoepidemiol Drug Saf 2025; 34:e70117. doi: 10.1002/pds.70117. [DOI] [PMC free article] [PubMed]
  • 93. Moon  RC, Bleak  TC, Rosenthal  NA, et al.  Epidemiology and economic burden of acute infectious gastroenteritis among adults treated in outpatient settings in US health systems. Am J Gastroenterol  2023; 118:1069–79. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94. Fitzke  H, Fayzan  T, Watkins  J, Galimov  E, Pierce  BF. Real-world evidence: state-of-the-art and future perspectives. J Comp Eff Res  2025; 14:e240130. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95. Garbi  M. National institute for health and care excellence clinical guidelines development principles and processes. Heart Br Card Soc  2021; 107:949–53. [DOI] [PubMed] [Google Scholar]
  • 96. Gargon  E, Williamson  PR, Altman  DG, Blazeby  JM, Tunis  S, Clarke  M. The COMET initiative database: progress and activities update (2015). Trials  2017; 18:54. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 97. Wang  SV, Pottegård  A, Crown  W, et al.  HARmonized protocol template to enhance reproducibility of hypothesis evaluating real-world evidence studies on treatment effects: a good practices report of a joint ISPE/ISPOR task force. Pharmacoepidemiol Drug Saf  2023; 32:44–55. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98. National Institute of Health (NIH) . Reimbursement Knowledge Guide for Diagnostics: NIH SEED Innovator Support Team 2023; Available at: https://seed.nih.gov/sites/default/files/2024-04/Reimbursement-Knowledge-Guide-for-Diagnostics.pdf. Accessed 7 March 2025.
  • 99.Applying Real-World Data and Real-World Evidence for Accelerated Approvals and Coverage Decisions. Available at: https://healthpolicy.duke.edu/publications/applying-real-world-data-and-real-world-evidence-accelerated-approvals-and-coverage. Accessed 19 March 2025.
  • 100. Lane-Fall  MB, Curran  GM, Beidas  RS. Scoping implementation science for the beginner: locating yourself on the ‘subway line’ of translational research. BMC Med Res Methodol  2019; 19:133. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 101. Broder-Fingert  S, Kuhn  J, Sheldrick  RC, et al.  Using the multiphase optimization strategy (MOST) framework to test intervention delivery strategies: a study protocol. Trials  2019; 20:728. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 102. Collins  LM, Murphy  SA, Strecher  V. The multiphase optimization strategy (MOST) and the sequential multiple assignment randomized trial (SMART): new methods for more potent eHealth interventions. Am J Prev Med  2007; 32:S112–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 103. Blonde  L, Khunti  K, Harris  SB, Meizinger  C, Skolnik  NS. Interpretation and impact of real-world clinical data for the practicing clinician. Adv Ther  2018; 35:1763–74. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 104. Hernán  MA, Wang  W, Leaf  DE. Target trial emulation: a framework for causal inference from observational data. JAMA  2022; 328:2446–7. [DOI] [PubMed] [Google Scholar]
  • 105. Fu  EL. Target trial emulation to improve causal inference from observational data: what, why, and how?  J Am Soc Nephrol  2023; 34:1305–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 106. Handley  MA, Lyles  CR, McCulloch  C, Cattamanchi  A. Selecting and improving quasi-experimental designs in effectiveness and implementation research. Annu Rev Public Health  2018; 39:5–25. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 107. Kim  Y, Steiner  P. Quasi-experimental designs for causal inference. Educ Psychol  2016; 51:395–405. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 108. Dimick  JB, Ryan  AM. Methods for evaluating changes in health care policy: the difference-in-differences approach. JAMA  2014; 312:2401–240. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

ofaf489_Supplementary_Data

Articles from Open Forum Infectious Diseases are provided here courtesy of Oxford University Press

RESOURCES