Abstract
Introduction
Outbreaks of communicable diseases in hospitals need to be quickly detected in order to enable immediate control. The increasing digitalization of hospital data processing offers potential solutions for automated outbreak detection systems (AODS). Our goal was to assess a newly developed AODS.
Methods
Our AODS was based on the diagnostic results of routine clinical microbiological examinations. The system prospectively counted detections per bacterial pathogen over time for the years 2016 and 2017. The baseline data covers data from 2013–2015. The comparative analysis was based on six different mathematical algorithms (normal/Poisson and score prediction intervals, the early aberration reporting system, negative binomial CUSUMs, and the Farrington algorithm). The clusters automatically detected were then compared with the results of our manual outbreak detection system.
Results
During the analysis period, 14 different hospital outbreaks were detected as a result of conventional manual outbreak detection. Based on the pathogens’ overall incidence, outbreaks were divided into two categories: outbreaks with rarely detected pathogens (sporadic) and outbreaks with often detected pathogens (endemic). For outbreaks with sporadic pathogens, the detection rate of our AODS ranged from 83% to 100%. Every algorithm detected 6 of 7 outbreaks with a sporadic pathogen. The AODS identified outbreaks with an endemic pathogen were at a detection rate of 33% to 100%. For endemic pathogens, the results varied based on the epidemiological characteristics of each outbreak and pathogen.
Conclusion
AODS for hospitals based on routine microbiological data is feasible and can provide relevant benefits for infection control teams. It offers in-time automated notification of suspected pathogen clusters especially for sporadically occurring pathogens. However, outbreaks of endemically detected pathogens need further individual pathogen-specific and setting-specific adjustments.
Introduction
Outbreak detection of infectious disease is an important area of hospital infection control. It is crucial to detect outbreaks as quickly as possible to limit, through early interventions, the potential for adverse outcomes in affected patients. Technical limitations pose a challenge for establishing a real-time detection system and early recognition of outbreaks [1–3]. In most cases, prospective outbreak detection relies on manual review of pooled microbiological results. This approach is currently being used successfully for rare multidrug-resistant organisms (MDRO)[3]. However, due to the higher number of susceptible organisms in comparison to MDROs, this approach not correlate with the expected outbreak risk in hospitals, but remains widely established as a result of an expected outbreak in hospitals, but remains widely established as a result of the high positive predictive value of very rare pathogens (e.g. with antimicrobial resistance) [4, 5]. The strength of this practical approach lies in the high positive predictive value of very rare pathogens (e.g. with antimicrobial resistance) [4, 5]. When limited human and laboratory resources are taken into consideration, the resulting low number of false positive results makes this outbreak detection manageable.
The increasing digitalization of hospital data offers increasing opportunities for prospective data analyses in modern hospitals [6]. This data can be used to systematically screen for unexpected increases in pathogen detection and thereby make automated outbreak detection systems possible, even for common pathogens [4, 7]. A challenge in finding valuable solutions for automated outbreak detection systems (AODS) is the comparability of the different analytic approaches [3]. In order to assess and compare, a universally agreed on definition of “outbreak” as a gold standard is needed, but is currently still lacking [8, 9].
In this work, our aim was to develop an AODS and to compare its results with our manual approach. Our AODS is based on mathematical methods and the source data was taken from real life hospital data.
Method
Our AODS is based on regular, computer-based, automated screening and systematic analyses of routinely collected microbiological laboratory and patient location data. In this work, we tested various methods of statistical analysis and compared them to the results of our current manual practice for outbreak detection.
Databases
The databases derived from real-time diagnostic results of the microbiology laboratory of Charité Universitätsmedizin Berlin. Taken as a whole, the hospital has more than 3,000 beds but is divided into three spatially separated hospital campuses. The individual campuses work mostly independently of each other and exchange patients only irregularly.
The period of the series of new outbreaks investigated in this paper ranged from 01/01/2016 until 12/31/2017. The manual detection of outbreaks occurred prospectively. Each outbreak investigated occurred during this time. The corresponding outbreak records were reported in such a way that the outbreak was at the mid-point of the datasets. Each set contained data for a 12 month period. Thus, it was possible to show data before and after the outbreak. Furthermore, the algorithms utilized by our AODS used historical databases from the same laboratory that cover a 3-year time span from 01/01/2013 and 31/12/2015. The historical databases were necessary to determine the baseline detection frequency of each pathogen. Only in this way could the AODS determine a suspicious change in pathogen detection frequency during the period in question. Data that included the following 11 species was analyzed: the Acinetobacter baumannii group (A. baumannii, A. pittii und A. nosocomialis), Citrobacter spp., Clostridium difficile, Enterococcus faecium, Enterobacter spp., Escherichia coli (only 3GCREB und CRE), Klebsiella spp., Pseudomonas aeruginosa, Salmonella spp., Serratia spp., and Staphylococcus aureus. Daily microbiological laboratory data was compiled over a 14-day period. A 14-day period was used because our physicians hold team meetings on a weekly basis to discuss the data from the previous 14 days. This period, which was set by the physicians, is based on transmission time. [10]
These pooled results were defined as a time interval.
The Institutional Review Board ‘Ethikkommission der Charité—Universitätsmedizin Berlin’ waived the requirement for data that is collected in alignment with the German Protection Against Infection Act. The data at hand serves explicitly for infection control purposes within the scope of this regulation.
Outbreak definitions
Established outbreak definition
In order to assess the results of our AODS, we compared the results to those of our manual outbreak detection system. The established system is comprised of pathogen-based daily manual review of bacteriological clinical and screening results as well as of information on infected patients collected by trained infection control physicians on their regular clinical rounds and is focused on MDROs. Whenever a suspicious increase of a certain pathogen is detected, our infection control physicians evaluate the likelihood of an outbreak. This begins with a review of the clinical and epidemiological data on the patients in question. If the likelihood of an outbreak remains high, further microbiological examinations are performed. This includes the collection of clinical microbiological material, the screening of patients in question as well as environmental examinations. A molecular biological assessment is later performed on the pathogens collected for clonality. If this is the case we consider an outbreak to be verified; if not the cluster is considered a false alarm. The outbreak investigation is conducted by at least one infection control resident and is supervised by an infection control attending physician. In addition, several infection control nurses provide basic clinical data and collect the microbiological samples.
AODS outbreak definition
For the analysis of the AODS, datasets were divided into endemic and sporadic pathogens, based on the frequency of detection per time interval. The criteria for “endemic” was met, when the pathogen was detected in more than 33% of analyzed time intervals. The definition “sporadic”, in contrast, was used for pathogens that were detected in less than 33% of analyzed time intervals. (Fig 1) An aberration was defined as a certain number of pathogens above the endemic level (marked as a colored bar in all figures). An endemic level for each database was determined by the algorithms mentioned above.
In order to develop an AODS with few false positive alarms, we specified a definition for ‘relevant outbreak alarm’ with a high positive predictive value (PPV). Every alarm was considered clinically relevant for datasets of sporadic pathogens. Such pathogens are rarely seen, hence relevant and false positive alarms are rare and no correction for specificity was necessary. Repeated detection of rare pathogens in different patients is, therefore, associated with ad hoc high positive predictive value. In order to ensure a timely alarm, the first (manually detected) outbreak time interval also needed to have triggered an AODS alarm.
Datasets of endemic pathogens are substantially more complicated. They are associated with a high likelihood of false positive alarms which would result in too much work in comparison to the additional benefit for the daily routine. In order to correct for this lack of specificity, we determined that a method is appropriate if 50% or less of the time intervals outside the outbreak were detected as aberrations. This eventually eliminates methods which have too many alarms. In addition, the timeliness of the alarm was ensured by an additional requirement. As with sporadic datasets, the first time interval of the manually detected outbreak needed to have triggered an AODS alarm. For datasets of endemic pathogens, only the combination of these two requirements led to the conclusion. AODS detected the same outbreak as the manual outbreak detection.
Algorithmic methods
There are five established categories of algorithms used to detect outbreaks. We utilized algorithms from three of these categories: prediction intervals, statistical process control, and statistical modeling. To predict intervals, we used normal distribution prediction intervals (PI-NV), Poisson distribution (PI-POI), and score prediction intervals (PI-SCORE) [11]. As methods of statistical process control, we used the early aberration reporting system (EARS) [12, 13] and negative binomial CUSUMs (NBC) [14]. For statistical modeling, we used the Farrington algorithm [15]. A more detailed description can be found in the supplement.
SatScan [16] based on temporal scan statistics was not used for several reasons. Our system was planned as a daily routine tool for infection control professionals. Furthermore, the technical requirements in our hospital led us to create an intranet application. Therefore, we created a graphical web user interface. At the time we developed our automated outbreak detection system, Satscan did not provide such a solution.
Previous publications have described the feasibility of machine learning for outbreak detection [17]. The method employed by Miller et al. required the measurement of pathogen similarity, e.g. whole genome sequencing. Unfortunately, this data was not routinely available in our hospital during the study period.
All 6 algorithms were used simultaneously. An aberration was detected if the number of isolates exceeded the threshold calculated by the 6 algorithms. Whether an outbreak was identified correctly depended on the pathogen (endemic or sporadic). A detected aberration was considered genuine, if it fulfilled the AODS outbreak definition from the section before. To compare the algorithm we calculated the “detection rate by the algorithms”, i.e. “correct identified outbreaks” divided by “all datasets” multiplied by 100.
All analyses were conducted in R [18]. The function “algo.farrington”was used from the package surveillance [19].
Results
Our infection control team detected 14 outbreaks in the two years analyzed (Table 1 and S1–S12 Figs). Seven outbreaks were found in datasets of pathogens classified as endemic (short endemic datasets), seven in datasets of sporadic pathogens (short sporadic datasets). The 14 outbreaks include six different bacterial species overall. The shortest outbreak duration was a single time interval (14 days), the longest 17 time intervals (238 days). In median, an endemic outbreak lasted 3 time intervals (mean: 6.14 time intervals) and a sporadic outbreak lasted in median 2 time intervals (mean: 2.00 time intervals).
Table 1. Overview of manually detected outbreaks in 2016 and 2017.
Outbreak | Pathogen | Drug Resistance | Start Time Interval1 of the outbreak | End Time Interval1 of the outbreak | Number of Isolates (involved in outbreak) | Time intervals with > = 1 isolates | Type of dataset |
---|---|---|---|---|---|---|---|
1 | Enterococcus faecium | VRE | 9 | 20 | 7 | 22 | Endemic2 |
2 | Enterococcus faecium | VRE | 9 | 18 | 17 | 25 | Endemic2 |
3 | Staphylococcus aureus | 13 | 14 | 6 | 24 | Endemic2 | |
4 | Clostridium difficile | 14 | 14 | 2 | 15 | Endemic2 | |
5 | Clostridium difficile | 14 | 14 | 2 | 15 | Endemic2 | |
6 | Enterococcus faecium | VRE | 14 | 16 | 10 | 22 | Endemic2 |
7 | Enterococcus faecium | VRE | 6 | 22 | 9 | 23 | Endemic2 |
I | Klebsiella spp | MDR | 13 | 14 | 3 | 4 | Sporadic3 |
II | Klebsiella spp | MDR | 14 | 16 | 6 | 6 | Sporadic3 |
III | Escherichia coli | XDR | 13 | 16 | 3 | 3 | Sporadic3 |
IV | Klebsiella spp | 14 | 14 | 8 | 7 | Sporadic3 | |
V | Acinetobacter baumannii | XDR | 13 | 15 | 3 | 5 | Sporadic3 |
VI | Clostridium difficile | 14 | 14 | 3 | 7 | Sporadic3 | |
VII | Clostridium difficile | 14 | 14 | 2 | 4 | Sporadic3 |
VRE, vancomycin-resistant enterococci. MDR, multidrug-resistant. XDR, extensively drug-resistant.
1Time interval equals 14 days.
2 Eendemic = Isolates found in more than 1/3 of time intervals investigated.
3 Ssporadic = Isolates found in 1/3 time intervals or less.
The respective courses of an endemic and sporadic pathogen are shown in Fig 2. Each bar represents the pathogens detected in one time interval. White bars indicate that no aberration was found, colored bars that an aberration was found. The AODS-calculated thresholds are shown as dotted lines in the figures. The blue boxes show the time frame of the outbreak by the manual method.
For the endemic dataset shown, a pathogen was found in almost every time interval (24 of 27). For the sporadic dataset at least one pathogen occurred in four time intervals.
The first bar in the outbreak of the endemic dataset was detected as an aberration by 5 of 6 algorithms. This met the first criterion for “outbreak was found” for an endemic dataset. For 3 of the 6 algorithms (POI-PI, SCORE-PI and Farrington algorithm) almost all bars outside the blue box indicate an aberration. This violates the second criterion for the outbreak definition of an endemic dataset. Hence, the endemic outbreak was detected by only two algorithms (NBC and NV-PI).
All endemic outbreaks were detected by normal distribution prediction interval (Table 2). Poisson prediction interval, score prediction interval, and negative binomial CUSUMs detected 6 of 7 outbreaks. Early aberration reporting system detected 5 of 7 outbreaks and the Farrington algorithm detected 1 of 7 outbreaks.
Table 2. Detection rate for endemic datasets, stratified by results from the 6 algorithms used.
Normal dirstribution prediction interval | Poisson dirstribution prediction interval | Score prediction interval | Early aberration reporting system | Negative Binomial Cusums | Farrington | Detection Rate of the outbreak | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
FF | L50 | FF | L50 | FF | L50 | FF | L50 | FF | L50 | FF | L50 | ||
Outbreak 1 (VRE) |
X | X | X | X | X | X | X | X | X | X | X | 83% | |
Outbreak 2 (VRE) |
X | X | X | X | X | X | X | X | 33% | ||||
Outbreak 3 (Staphylo-coccus aureus) |
X | X | X | X | X | X | X | X | X | X | X | X | 100% |
Outbreak 4 (CDIF) |
X | X | X | X | X | X | X | 50% | |||||
Outbreak 5 (CDIF) |
X | X | X | X | X | X | X | X | X | X | X | 83% | |
Outbreak 6 (VRE) |
X | X | X | X | X | X | X | X | X | X | X | 83% | |
Outbreak 7 (VRE) |
X | X | X | X | X | X | X | X | X | X | 83% | ||
Detection rate of the alorithms | 100% | 86% | 86% | 71% | 86% | 14% |
FF (First Found), first outbreak time interval detected as aberration. L50, ≤50% of the time intervals outside the outbreak were detected as aberration. X, the condition FF or L50 was met. An outbreak detection required that both conditions be met. Coloured background indicates that the outbreak was detected by our Automated Outbreak Detection System. VRE Vancomycin resitant Enterococci. CDIF Clostridum difficile.
Almost every first time interval in an endemic outbreak was detected, except in 4 cases. In 9 of 42 cases of endemic outbreak, the algorithms detected too many false alarms. Six of them resulted from use of the Farrington algorithm.
Every algorithm except negative binomial CUSUMs detected all sporadic outbreaks (Table 3).
Table 3. Detection rate for sporadic datasets, stratified by results from the 6 algorithms used.
Normal dirstribution prediction interval | Poisson dirstribution prediction interval | Score prediction interval | Early aberration reporting system | Negative Binomial Cusums | Farrington | Detection Rate of the outbreak | |
---|---|---|---|---|---|---|---|
FF | FF | FF | FF | FF | FF | ||
Outbreak I (MDR Klebsiella spp.) |
X | X | X | X | X | X | 100% |
Outbreak II (MDR Klebsiella spp.) |
X | X | X | X | X | X | 100% |
Outbreak III (XDR Escherichia coli) |
X | X | X | X | X | 83% | |
Outbreak IV (Klebsiella spp.) |
X | X | X | X | X | X | 100% |
Outbreak V (XDR Acinobacter baumanii) |
X | X | X | X | X | 83% | |
Outbreak VI (CDIF) |
X | X | X | X | X | X | 100% |
Outbreak VII (CDIF) |
X | X | X | X | X | X | 100% |
Detection rate of the alorithms | 100% | 100% | 100% | 100% | 71% | 100% |
FF (First Found), first outbreak time interval was detected as aberration. L50, ≤50% of the time intervals outside the outbreak were detected as aberration. X, the condition FF was met. Coloured background indicates that the outbreak was detected by our Automated Outbreak Detection System. MDR multidrug resistant. XDR extensivily drug resistant. CDIF clostridium difficile.
Discussion
When analyzing routine diagnostic microbiological data, automated outbreak detection systems (AODS) offer the means of screening for outbreaks of both common as well as sporadic pathogens[4]. In this work, our aim was to assess an AODS we developed for our hospital. The main approach was to compare our established manual outbreak detection system with the AODS under real-life conditions.
Sporadic vs. endemic
Based on a rough epidemiological pattern, we defined two types of outbreak analyses: one for pathogens detected sporadically [7, 11] and one for pathogens with high frequency of detection. We termed such pathogens endemic pathogens. These pose a relevant detection problem as they often produce false positive alarms leading to a high risk of pseudo-outbreaks [20]. In contrast, in the case of sporadically detected pathogens, the positive predictive value of an outbreak alarm is very high. If very rare pathogens are detected in multiple patients in a closed space (e.g. particular ward) within a short time (e.g. one month), an epidemiological correlation is highly likely. In this case, no complicated algorithmic analysis is necessary. The benefit of an automated system, then, lies in its immediate alarm. With endemic pathogens, outbreak investigations require close cooperation with the respective clinical department. Microbiological detection of these common pathogens rarely triggers suspicion of an outbreak based on the recurrence of the pathogen. However, it must be assumed that most outbreaks occur with common pathogens. Hence, it is very likely that there is a large reporting bias favoring outbreaks with rare or MDR pathogens [21]. Highly discriminatory molecular analytical methods, e.g. whole genome sequencing, offer a possible future solution since pathogens can be quickly evaluated if suspicious clusters occur [22]. Unfortunately, for the time being these methods are time consuming, expensive, and require specialists trained in bioinformatics. Therefore, they are not currently available as routine diagnostic methods at most hospitals and often are applied only in situations where there exists high risks of further transmission or to patient health.
Outbreaks of sporadically detected pathogens
All outbreaks with sporadic pathogens were detected by 5 of 6 algorithms. NBC missed an XDR E. coli- and XDR A. baumannii outbreak. In order to provide the infection control team with a real benefit in terms of reaction time, our definition of detection required an alarm at the beginning of the outbreak. Even though the XDR A. baumannii outbreak was eventually detected, it happened only after a significant delay of more than one month (S4 Fig). The XDR E. coli outbreak was not detected by NBC at all (S2 Fig). In contrast to our results, Watkins et al. showed that NBC provides better detection rates and fewer false alarms than others, e.g. EARS [13]. Another simulation study demonstrated the superiority of NBC to EARS [23]. Those studies, however, tested NBC for natural outbreaks of disease syndromes with a high number of patients, as in natural outbreaks of viral diseases[13]. It is questionable if the circumstances of these studies are can be compared to our setting (hospital ward, bacterial pathogen, low number of cases). Further studies are needed to assess the detection detection rates of these algorithms, in particular NBC, for outbreaks in hospital wards with sporadic bacteria.
Outbreaks of endemic pathogens
Concerning outbreaks of endemic pathogens, the algorithm’s ‘normal distribution prediction interval’ was superior to all others and detected every outbreak, whereas Farrington performed worst.
Despite Farrington’s detection of 6 of 7 outbreaks, it produced too many false alarms. Our results concerning the Farrington algorithm stand in contrast to the published literature which report good performance by regression models like Farrington, e.g. in public health in France [24]. However, Farrington was designed to adjust for seasonality, which does not apply in our datasets. Moreover, regression models like Farrington do not perform well for small outbreaks (low number of patients) like those in our hospital and in most other hospital settings [25]. Previously it was shown in a simulation study that Farrington has a low sensitivity [26]. Therefore, it is possible that the Farrington algorithm alone should not be the algorithm of choice for regular hospital outbreaks.
Poisson distribution, score distribution prediction interval as well as NBC performed similarly well and showed weaknesses only regarding a VR-E. faecium outbreak. In comparison, EARS missed one additional outbreak, one of two C. difficile outbreaks. The VR-E. faecium outbreak missed occurred during an increase in VR-E. faecium incidence that was slow overall. This most likely led to an increase in false alarms in Poisson distribution and score distribution. NBC, however, reacted to this increase with fewer false alarms (higher specificity) but triggered the alarm with an unacceptable delay of more than 1.5 months after the outbreak onset.
Moreover, it should be acknowledged that in several respects, outbreaks with VRE are outstanding examples of infection control practice. First, VR-E. faecium is currently a very common pathogen with significantly increasing incidence [27] in Germany. In our datasets comprise four different outbreaks which represent the highest prevalence (up to 14 different patients per time period). Second, different VRE isolates from hospital clusters are often not distinguishable using commonly available molecular methods [28]. Moreover, because enterococci are ubiquitous bacteria, transmission routes are highly complex and include various modes of introduction between different hospitals and wards, especially oncology wards with continually returning VRE patients [28, 29]. This eventually leads to low positive predictive value for detected aberrations of clusters on oncology wards (or other high endemic settings). This is different particularly in other disciplines with low VRE prevalence. Therefore, adjusted levels of specificity based on ward discipline are required, notably for VRE in Germany, but possibly also for pathogens with similar epidemiological characteristics.
Outbreaks of Clostridium difficile infections (CDI)
Outbreaks of toxin-producing C. difficile are another problematic area of automated outbreak detection. C. difficile is a ubiquitous bacterium; the associated diarrhea (CDI) usually follows extensive antimicrobial therapy [30]. Therefore, simple pathogen detection without clinical verification of diarrhea is most likely not an adequate approach for CDI outbreak detection. Moreover, whole-genome outbreak studies showed that probably only 40% of nosocomial CDI clusters are clonal [31]. The rest are most probably independent events that follow antimicrobial therapy. They can appear as clusters due to ward-specific policies determining which antimicrobial therapies are used. Therefore, on a high endemic ward, the positive predictive value of a CDI outbreak alarm is low. Even more problematic, most laboratories examine factors other than bacterial cultures, for example toxin production. This often renders it impossible to determine post hoc clonal relatedness. We examined four different C. difficile clusters. Two of them were in sporadic situations and were detected by all algorithms (detection rate 100%). Of the two other endemic clusters, one was detected with an acceptable detection rate of 83% (outbreak 5). The other one (outbreak 4) was detected by only 50% of the algorithms we tested. A noteworthy difference between these outbreaks was the steeper increase in detection rates during outbreak 5 (compared to outbreak 4) and the associated period prior the outbreak. Our results, therefore, showed that for CDI further adjustments need to be made, in order determine the likelihood of a CDI outbreak.
Limitations
Our AODS approach is pathogen-based and can only detect possible clonal outbreaks. No gold standard for outbreak detection exists because there is no international consensus on the definition of an outbreak [3, 4, 20]. Therefore, specificity could not be calculated for our AODS. Hence, we have based our analysis on the comparison of the outbreak detection system we currently use and the AODS. Our manual outbreak detection system works prospectively but the AODS works retrospectively. This approach could have missed outbreaks. Hence, such outbreaks could be not analyzed by AODS. These Another limitation exists with regard to clonal typing which was not performed routinely. Therefore, it was not possible to detect the potential for better detection rates of the AODS with respect to our manual outbreak detection. Future prospective studies that include state-of-the-art clonal typing methods are necessary in order to assess further benefits of AODS.
Conclusion
Our work showed that an automated outbreak system for sporadic bacteria in hospitals can work reliably in many cases. It can provide an early warning system and depends only on timely reports of microbiological results. The greatest benefit of such an automated system lies in the automatic alarm for clusters of otherwise rare pathogens, especially in large hospitals. Regarding more common bacteria, the system resulted in a substantial improvement in one hospital’s outbreak detection detection rates. However, the low positive predictive value of those alarms illustrates the need for further adjustments for various other variables. For example, ward and pathogen-specific characteristics need to be taken into consideration in all analyses because they change the predictive value to a great extent. If outbreak isolates are still retrievable, the alarm would lead to further molecular analyses of the isolates’ genetic relatedness. In most cases, this will not be feasible and a decision on the likelihood of a real outbreak remains dependent on the expertise of the hospital epidemiologist. We believe that currently available technology cannot replace an experienced hospital epidemiologist. However, although it needs further development and evaluation in real life situations, our system provides fundamental work toward a system of automated outbreak detection.
Supporting information
Acknowledgments
We wish to thank our infection control physicians for their daily routine work on our system and their valuable feedback. These are: Peter Bischoff (MD), Christine Geffers (MD), Sonja Hansen (MD), Julia Hermes (MD), Axel Kola (MD), Tobias Kramer (MD), Friederike Maechler (MD), Cornelius Remschmidt (MD), Florian Salm (MD), Beate Weikert (MD), Miriam Wiese-Posselt (MD).
Special thanks go also to our infection control nurses.
Thanks to Gerald Brennan for proofreading.
Abbreviations
- Aberration
AODS notification that the number of detected pathogens per time interval is above the expected detection frequency regardless whether an outbreak has been confirmed
- Algorithm
Statistical application that evaluates the likelihood of an outbreak behind fluctuating detection frequencies of pathogens based on the comparison to historical data
- AODS
Automated outbreak detection system
- CDI
Clostridium difficile infection
- Cluster
An accumulation of phenotypically identical pathogens within a narrow area and time frame
- EARS
Early aberration reporting system
- NBC
Negative binomial CUSUMs
- Outbreak (manual)
A cluster of patients with the phenotypically identical pathogen that was considered to derive from nosocomial transmission based on the judgment of the manual outbreak detection system
- PI-NV
Normal distribution prediction intervals
- PI-POI
Poisson distributed prediction intervals
- Time interval
14 days of microbiological results that were analyzed as a single period
- VRE
Vancomycin-resistant enterococci
Data Availability
All relevant data are available from https://zenodo.org/record/3603309.
Funding Statement
This work was partly funded by grants from the German Society of Hospital Hygiene (DGKH). The concept of the project won a scientific price from Stiftung Charité. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Snitkin ES, Zelazny AM, Thomas PJ, Stock F, Henderson DK, Palmore TN, et al. Tracking a hospital outbreak of carbapenem-resistant Klebsiella pneumoniae with whole-genome sequencing. Science translational medicine. 2012;4(148):148ra16–ra16. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Bergin SM, Periaswamy B, Barkham T, Chua HC, Mok YM, Fung DSS, et al. An Outbreak of Streptococcus pyogenes in a Mental Health Facility: Advantage of Well-Timed Whole-Genome Sequencing Over emm Typing. Infection control and hospital epidemiology. 2018:1–9. [DOI] [PubMed] [Google Scholar]
- 3.Baker MA, Huang SS, Letourneau AR, Kaganov RE, Peeples JR, Drees M, et al. Lack of comprehensive outbreak detection in hospitals. infection control & hospital epidemiology. 2016;37(4):466–8. [DOI] [PubMed] [Google Scholar]
- 4.Leclere B, Buckeridge DL, Boelle PY, Astagneau P, Lepelletier D. Automated detection of hospital outbreaks: A systematic review of methods. PLoS One. 2017;12(4):e0176438 10.1371/journal.pone.0176438 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Buckeridge DL. Outbreak detection through automated surveillance: a review of the determinants of detection. Journal of biomedical informatics. 2007;40(4):370–9. 10.1016/j.jbi.2006.09.003 [DOI] [PubMed] [Google Scholar]
- 6.Murdoch TB, Detsky AS. The inevitable application of big data to health care. Jama. 2013;309(13):1351–2. 10.1001/jama.2013.393 [DOI] [PubMed] [Google Scholar]
- 7.Dessau RB, Steenberg P. Computerized surveillance in clinical microbiology with time series analysis. Journal of clinical microbiology. 1993;31(4):857–60. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Buehler JW, Hopkins RS, Overhage JM, Sosin DM, Tong V. Framework for evaluating public health surveillance systems for early detection of outbreaks: recommendations from the CDC Working Group. MMWR Recommendations and reports: Morbidity and mortality weekly report Recommendations and reports. 2004;53(Rr-5):1–11. [PubMed] [Google Scholar]
- 9.Rutjes AW, Reitsma JB, Coomarasamy A, Khan KS, Bossuyt PM. Evaluation of diagnostic tests when there is no gold standard. A review of methods. Health technology assessment (Winchester, England). 2007;11(50):iii, ix–51. [DOI] [PubMed] [Google Scholar]
- 10.Grundmann H, Barwolff S, Tami A, Behnke M, Schwab F, Geffers C, et al. How many infections are caused by patient-to-patient transmission in intensive care units? Critical care medicine. 2005;33(5):946–51. 10.1097/01.ccm.0000163223.26234.56 [DOI] [PubMed] [Google Scholar]
- 11.Nishiura H. Early detection of nosocomial outbreaks caused by rare pathogens: a case study employing score prediction interval. Osong Public Health Res Perspect. 2012;3(3):121–7. 10.1016/j.phrp.2012.07.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Hutwagner L, Thompson W, Seeman GM, Treadwell T. The bioterrorism preparedness and response Early Aberration Reporting System (EARS). J Urban Health. 2003;80(2 Suppl 1):i89–96. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Watkins RE, Eagleson S, Veenendaal B, Wright G, Plant AJ. Applying cusum-based methods for the detection of outbreaks of Ross River virus disease in Western Australia. BMC Med Inform Decis Mak. 2008;8:37 10.1186/1472-6947-8-37 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Pelecanos AM, Ryan PA, Gatton ML. Outbreak detection algorithms for seasonal disease data: a case study using Ross River virus disease. BMC Med Inform Decis Mak. 2010;10:74 10.1186/1472-6947-10-74 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Farrington CP, Andrews NJ, Beale AD, Catchpole MA. A Statistical Algorithm for the Early Detection of Outbreaks of Infectious Disease. Journal of the Royal Statistical Society Series A (Statistics in Society). 1996;159(3):547–63. [Google Scholar]
- 16.Kulldorff M, Heffernan R, Hartman J, Assuncao R, Mostashari F. A space-time permutation scan statistic for disease outbreak detection. PLoS medicine. 2005;2(3):e59 10.1371/journal.pmed.0020059 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Miller JK, Chen J, Sundermann A, Marsh JW, Saul MI, Shutt KA, et al. Statistical outbreak detection by joining medical records and pathogen similarity. J Biomed Inform. 2019;91:103126 10.1016/j.jbi.2019.103126 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.R Development Core Team. R: A language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing; 2019. [Google Scholar]
- 19.Salmon M, Schumacher D, Hohle M. Monitoring Count Time Series in R: Aberration Detection in Public Health Surveillance. J Stat Softw. 2016;70(10):1–35. [Google Scholar]
- 20.Stoesser N, Sheppard AE, Moore C, Golubchik T, Parry C, Nget P, et al. Extensive within-host diversity in fecally carried extended-spectrum beta-lactamase-producing Escherichia coli: implications for transmission analyses. Journal of clinical microbiology. 2015:JCM. 00378–15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Gastmeier P, Stamm-Balderjahn S, Hansen S, Zuschneid I, Sohr D, Behnke M, et al. Where should one search when confronted with outbreaks of nosocomial infection? American journal of infection control. 2006;34(9):603–5. 10.1016/j.ajic.2006.01.014 [DOI] [PubMed] [Google Scholar]
- 22.Sabat AJ, Budimir A, Nashev D, Sá-Leão R, Van Dijl JM, Laurent F, et al. Overview of molecular typing methods for outbreak detection and epidemiological surveillance. Eurosurveillance. 2013;18(4):20380 10.2807/ese.18.04.20380-en [DOI] [PubMed] [Google Scholar]
- 23.Fricker RD, Hegler BL Jr., Dunfee DA. Comparing syndromic surveillance detection methods: EARS' versus a CUSUM-based methodology. Stat Med. 2008;27(17):3407–29. 10.1002/sim.3197 [DOI] [PubMed] [Google Scholar]
- 24.Guillou A, Kratz M, Strat YL. An extreme value theory approach for the early detection of time clusters. A simulation-based assessment and an illustration to the surveillance of Salmonella. Statistics in medicine. 2014;33(28):5015–27. 10.1002/sim.6275 [DOI] [PubMed] [Google Scholar]
- 25.Unkel S, Farrington C, Garthwaite PH, Robertson C, Andrews N. Statistical methods for the prospective detection of infectious disease outbreaks: a review. Journal of the Royal Statistical Society: Series A (Statistics in Society). 2012;175(1):49–82. [Google Scholar]
- 26.Bedubourg G, Le Strat Y. Evaluation and comparison of statistical methods for early temporal detection of outbreaks: A simulation-based study. PLoS One. 2017;12(7):e0181227 10.1371/journal.pone.0181227 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Remschmidt C, Schroder C, Behnke M, Gastmeier P, Geffers C, Kramer TS. Continuous increase of vancomycin resistance in enterococci causing nosocomial infections in Germany—10 years of surveillance. Antimicrobial resistance and infection control. 2018;7:54 10.1186/s13756-018-0353-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Raven KE, Gouliouris T, Brodrick H, Coll F, Brown NM, Reynolds R, et al. Complex routes of nosocomial vancomycin-resistant Enterococcus faecium transmission revealed by genome sequencing. Clinical infectious diseases. 2017;64(7):886–93. 10.1093/cid/ciw872 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Ulrich N, Vonberg RP, Gastmeier P. Outbreaks caused by vancomycin-resistant Enterococcus faecium in hematology and oncology departments: A systematic review. Heliyon. 2017;3(12):e00473 10.1016/j.heliyon.2017.e00473 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Leffler DA, Lamont JT. Clostridium difficile infection. New England Journal of Medicine. 2015;372(16):1539–48. 10.1056/NEJMra1403772 [DOI] [PubMed] [Google Scholar]
- 31.Eyre DW, Cule ML, Walker AS, Crook DW, Wilcox MH, Peto TE. Hospital and community transmission of Clostridium difficile: a whole genome sequencing study. The Lancet. 2012;380:S12. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
All relevant data are available from https://zenodo.org/record/3603309.