#PAGE_PARAMS# #ADS_HEAD_SCRIPTS# #MICRODATA#

Lean back and wait for the alarm? Testing an automated alarm system for nosocomial outbreaks to provide support for infection control professionals


Authors: Christin Schröder aff001;  Luis Alberto Peña Diaz aff001;  Anna Maria Rohde aff001;  Brar Piening aff001;  Seven Johannes Sam Aghdassi aff001;  Georg Pilarski aff001;  Norbert Thoma aff001;  Petra Gastmeier aff001;  Rasmus Leistner aff001;  Michael Behnke aff001
Authors place of work: Charité – Universitätsmedizin Berlin, Institute of Hygiene and Environmental Medicine, Berlin, Germany aff001
Published in the journal: PLoS ONE 15(1)
Category: Research Article
doi: https://doi.org/10.1371/journal.pone.0227955

Summary

Introduction

Outbreaks of communicable diseases in hospitals need to be quickly detected in order to enable immediate control. The increasing digitalization of hospital data processing offers potential solutions for automated outbreak detection systems (AODS). Our goal was to assess a newly developed AODS.

Methods

Our AODS was based on the diagnostic results of routine clinical microbiological examinations. The system prospectively counted detections per bacterial pathogen over time for the years 2016 and 2017. The baseline data covers data from 2013–2015. The comparative analysis was based on six different mathematical algorithms (normal/Poisson and score prediction intervals, the early aberration reporting system, negative binomial CUSUMs, and the Farrington algorithm). The clusters automatically detected were then compared with the results of our manual outbreak detection system.

Results

During the analysis period, 14 different hospital outbreaks were detected as a result of conventional manual outbreak detection. Based on the pathogens’ overall incidence, outbreaks were divided into two categories: outbreaks with rarely detected pathogens (sporadic) and outbreaks with often detected pathogens (endemic). For outbreaks with sporadic pathogens, the detection rate of our AODS ranged from 83% to 100%. Every algorithm detected 6 of 7 outbreaks with a sporadic pathogen. The AODS identified outbreaks with an endemic pathogen were at a detection rate of 33% to 100%. For endemic pathogens, the results varied based on the epidemiological characteristics of each outbreak and pathogen.

Conclusion

AODS for hospitals based on routine microbiological data is feasible and can provide relevant benefits for infection control teams. It offers in-time automated notification of suspected pathogen clusters especially for sporadically occurring pathogens. However, outbreaks of endemically detected pathogens need further individual pathogen-specific and setting-specific adjustments.

Keywords:

Clostridium difficile – Bacterial pathogens – Pathogens – Algorithms – Infectious disease control – Toxins – Epidemiology – Binomials

Introduction

Outbreak detection of infectious disease is an important area of hospital infection control. It is crucial to detect outbreaks as quickly as possible to limit, through early interventions, the potential for adverse outcomes in affected patients. Technical limitations pose a challenge for establishing a real-time detection system and early recognition of outbreaks [13]. In most cases, prospective outbreak detection relies on manual review of pooled microbiological results. This approach is currently being used successfully for rare multidrug-resistant organisms (MDRO)[3]. However, due to the higher number of susceptible organisms in comparison to MDROs, this approach not correlate with the expected outbreak risk in hospitals, but remains widely established as a result of an expected outbreak in hospitals, but remains widely established as a result of the high positive predictive value of very rare pathogens (e.g. with antimicrobial resistance) [4, 5]. The strength of this practical approach lies in the high positive predictive value of very rare pathogens (e.g. with antimicrobial resistance) [4, 5]. When limited human and laboratory resources are taken into consideration, the resulting low number of false positive results makes this outbreak detection manageable.

The increasing digitalization of hospital data offers increasing opportunities for prospective data analyses in modern hospitals [6]. This data can be used to systematically screen for unexpected increases in pathogen detection and thereby make automated outbreak detection systems possible, even for common pathogens [4, 7]. A challenge in finding valuable solutions for automated outbreak detection systems (AODS) is the comparability of the different analytic approaches [3]. In order to assess and compare, a universally agreed on definition of “outbreak” as a gold standard is needed, but is currently still lacking [8, 9].

In this work, our aim was to develop an AODS and to compare its results with our manual approach. Our AODS is based on mathematical methods and the source data was taken from real life hospital data.

Method

Our AODS is based on regular, computer-based, automated screening and systematic analyses of routinely collected microbiological laboratory and patient location data. In this work, we tested various methods of statistical analysis and compared them to the results of our current manual practice for outbreak detection.

Databases

The databases derived from real-time diagnostic results of the microbiology laboratory of Charité Universitätsmedizin Berlin. Taken as a whole, the hospital has more than 3,000 beds but is divided into three spatially separated hospital campuses. The individual campuses work mostly independently of each other and exchange patients only irregularly.

The period of the series of new outbreaks investigated in this paper ranged from 01/01/2016 until 12/31/2017. The manual detection of outbreaks occurred prospectively. Each outbreak investigated occurred during this time. The corresponding outbreak records were reported in such a way that the outbreak was at the mid-point of the datasets. Each set contained data for a 12 month period. Thus, it was possible to show data before and after the outbreak. Furthermore, the algorithms utilized by our AODS used historical databases from the same laboratory that cover a 3-year time span from 01/01/2013 and 31/12/2015. The historical databases were necessary to determine the baseline detection frequency of each pathogen. Only in this way could the AODS determine a suspicious change in pathogen detection frequency during the period in question. Data that included the following 11 species was analyzed: the Acinetobacter baumannii group (A. baumannii, A. pittii und A. nosocomialis), Citrobacter spp., Clostridium difficile, Enterococcus faecium, Enterobacter spp., Escherichia coli (only 3GCREB und CRE), Klebsiella spp., Pseudomonas aeruginosa, Salmonella spp., Serratia spp., and Staphylococcus aureus. Daily microbiological laboratory data was compiled over a 14-day period. A 14-day period was used because our physicians hold team meetings on a weekly basis to discuss the data from the previous 14 days. This period, which was set by the physicians, is based on transmission time. [10]

These pooled results were defined as a time interval.

The Institutional Review Board ‘Ethikkommission der Charité—Universitätsmedizin Berlin’ waived the requirement for data that is collected in alignment with the German Protection Against Infection Act. The data at hand serves explicitly for infection control purposes within the scope of this regulation.

Outbreak definitions

Established outbreak definition

In order to assess the results of our AODS, we compared the results to those of our manual outbreak detection system. The established system is comprised of pathogen-based daily manual review of bacteriological clinical and screening results as well as of information on infected patients collected by trained infection control physicians on their regular clinical rounds and is focused on MDROs. Whenever a suspicious increase of a certain pathogen is detected, our infection control physicians evaluate the likelihood of an outbreak. This begins with a review of the clinical and epidemiological data on the patients in question. If the likelihood of an outbreak remains high, further microbiological examinations are performed. This includes the collection of clinical microbiological material, the screening of patients in question as well as environmental examinations. A molecular biological assessment is later performed on the pathogens collected for clonality. If this is the case we consider an outbreak to be verified; if not the cluster is considered a false alarm. The outbreak investigation is conducted by at least one infection control resident and is supervised by an infection control attending physician. In addition, several infection control nurses provide basic clinical data and collect the microbiological samples.

AODS outbreak definition

For the analysis of the AODS, datasets were divided into endemic and sporadic pathogens, based on the frequency of detection per time interval. The criteria for “endemic” was met, when the pathogen was detected in more than 33% of analyzed time intervals. The definition “sporadic”, in contrast, was used for pathogens that were detected in less than 33% of analyzed time intervals. (Fig 1) An aberration was defined as a certain number of pathogens above the endemic level (marked as a colored bar in all figures). An endemic level for each database was determined by the algorithms mentioned above.

Classification of outbreaks into two types.
Fig. 1. Classification of outbreaks into two types.
Datasets with endemically detected pathogens and datasets with sporadic pathogens. In the endemic dataset at least one pathogen occurred more than 30% of the time. In the sporadic dataset at least one pathogen occurred 30% or less of the time.

In order to develop an AODS with few false positive alarms, we specified a definition for ‘relevant outbreak alarm’ with a high positive predictive value (PPV). Every alarm was considered clinically relevant for datasets of sporadic pathogens. Such pathogens are rarely seen, hence relevant and false positive alarms are rare and no correction for specificity was necessary. Repeated detection of rare pathogens in different patients is, therefore, associated with ad hoc high positive predictive value. In order to ensure a timely alarm, the first (manually detected) outbreak time interval also needed to have triggered an AODS alarm.

Datasets of endemic pathogens are substantially more complicated. They are associated with a high likelihood of false positive alarms which would result in too much work in comparison to the additional benefit for the daily routine. In order to correct for this lack of specificity, we determined that a method is appropriate if 50% or less of the time intervals outside the outbreak were detected as aberrations. This eventually eliminates methods which have too many alarms. In addition, the timeliness of the alarm was ensured by an additional requirement. As with sporadic datasets, the first time interval of the manually detected outbreak needed to have triggered an AODS alarm. For datasets of endemic pathogens, only the combination of these two requirements led to the conclusion. AODS detected the same outbreak as the manual outbreak detection.

Algorithmic methods

There are five established categories of algorithms used to detect outbreaks. We utilized algorithms from three of these categories: prediction intervals, statistical process control, and statistical modeling. To predict intervals, we used normal distribution prediction intervals (PI-NV), Poisson distribution (PI-POI), and score prediction intervals (PI-SCORE) [11]. As methods of statistical process control, we used the early aberration reporting system (EARS) [12, 13] and negative binomial CUSUMs (NBC) [14]. For statistical modeling, we used the Farrington algorithm [15]. A more detailed description can be found in the supplement.

SatScan [16] based on temporal scan statistics was not used for several reasons. Our system was planned as a daily routine tool for infection control professionals. Furthermore, the technical requirements in our hospital led us to create an intranet application. Therefore, we created a graphical web user interface. At the time we developed our automated outbreak detection system, Satscan did not provide such a solution.

Previous publications have described the feasibility of machine learning for outbreak detection [17]. The method employed by Miller et al. required the measurement of pathogen similarity, e.g. whole genome sequencing. Unfortunately, this data was not routinely available in our hospital during the study period.

All 6 algorithms were used simultaneously. An aberration was detected if the number of isolates exceeded the threshold calculated by the 6 algorithms. Whether an outbreak was identified correctly depended on the pathogen (endemic or sporadic). A detected aberration was considered genuine, if it fulfilled the AODS outbreak definition from the section before. To compare the algorithm we calculated the “detection rate by the algorithms”, i.e. “correct identified outbreaks” divided by “all datasets” multiplied by 100.

All analyses were conducted in R [18]. The function “algo.farrington”was used from the package surveillance [19].

Results

Our infection control team detected 14 outbreaks in the two years analyzed (Table 1 and S1S12 Figs). Seven outbreaks were found in datasets of pathogens classified as endemic (short endemic datasets), seven in datasets of sporadic pathogens (short sporadic datasets). The 14 outbreaks include six different bacterial species overall. The shortest outbreak duration was a single time interval (14 days), the longest 17 time intervals (238 days). In median, an endemic outbreak lasted 3 time intervals (mean: 6.14 time intervals) and a sporadic outbreak lasted in median 2 time intervals (mean: 2.00 time intervals).

Tab. 1. Overview of manually detected outbreaks in 2016 and 2017.
Overview of manually detected outbreaks in 2016 and 2017.
Endemic outbreaks are indicated by Arabic numerals, sporadic outbreaks by Roman numerals.

The respective courses of an endemic and sporadic pathogen are shown in Fig 2. Each bar represents the pathogens detected in one time interval. White bars indicate that no aberration was found, colored bars that an aberration was found. The AODS-calculated thresholds are shown as dotted lines in the figures. The blue boxes show the time frame of the outbreak by the manual method.

Two examples of outbreaks detected manually vs. outbreaks detected by AODS.
Fig. 2. Two examples of outbreaks detected manually vs. outbreaks detected by AODS.
Left: Outbreak in an endemic dataset (outbreak 2, vancomycin-resistant E. faecium). Right: Outbreak in a sporadic dataset (outbreak I, Klebisella spp., MDR). Depicted is the course of pathogen detection on the ward during a year when an outbreak was manually detected. The manually detected outbreak is in the center and is indicated by a light blue box. Each bar represents the number of pathogens detected per time interval (14 days). If a bar is colored, an algorithm detected an aberration. Shown are the results for all six algorithms (top down in different colors): normal prediction interval, Poisson prediction interval, score prediction interval, early aberration report system, negative binomial CUSUMs, and the Farrington algorithm.

For the endemic dataset shown, a pathogen was found in almost every time interval (24 of 27). For the sporadic dataset at least one pathogen occurred in four time intervals.

The first bar in the outbreak of the endemic dataset was detected as an aberration by 5 of 6 algorithms. This met the first criterion for “outbreak was found” for an endemic dataset. For 3 of the 6 algorithms (POI-PI, SCORE-PI and Farrington algorithm) almost all bars outside the blue box indicate an aberration. This violates the second criterion for the outbreak definition of an endemic dataset. Hence, the endemic outbreak was detected by only two algorithms (NBC and NV-PI).

All endemic outbreaks were detected by normal distribution prediction interval (Table 2). Poisson prediction interval, score prediction interval, and negative binomial CUSUMs detected 6 of 7 outbreaks. Early aberration reporting system detected 5 of 7 outbreaks and the Farrington algorithm detected 1 of 7 outbreaks.

Tab. 2. Detection rate for endemic datasets, stratified by results from the 6 algorithms used.
Detection rate for endemic datasets, stratified by results from the 6 algorithms used.
The detection rate is shown for each algorithm (columns) and each outbreak (rows).

Almost every first time interval in an endemic outbreak was detected, except in 4 cases. In 9 of 42 cases of endemic outbreak, the algorithms detected too many false alarms. Six of them resulted from use of the Farrington algorithm.

Every algorithm except negative binomial CUSUMs detected all sporadic outbreaks (Table 3).

Tab. 3. Detection rate for sporadic datasets, stratified by results from the 6 algorithms used.
Detection rate for sporadic datasets, stratified by results from the 6 algorithms used.
The detection rate is shown for each algorithm (columns) and each outbreak (rows).

Discussion

When analyzing routine diagnostic microbiological data, automated outbreak detection systems (AODS) offer the means of screening for outbreaks of both common as well as sporadic pathogens[4]. In this work, our aim was to assess an AODS we developed for our hospital. The main approach was to compare our established manual outbreak detection system with the AODS under real-life conditions.

Sporadic vs. endemic

Based on a rough epidemiological pattern, we defined two types of outbreak analyses: one for pathogens detected sporadically [7, 11] and one for pathogens with high frequency of detection. We termed such pathogens endemic pathogens. These pose a relevant detection problem as they often produce false positive alarms leading to a high risk of pseudo-outbreaks [20]. In contrast, in the case of sporadically detected pathogens, the positive predictive value of an outbreak alarm is very high. If very rare pathogens are detected in multiple patients in a closed space (e.g. particular ward) within a short time (e.g. one month), an epidemiological correlation is highly likely. In this case, no complicated algorithmic analysis is necessary. The benefit of an automated system, then, lies in its immediate alarm. With endemic pathogens, outbreak investigations require close cooperation with the respective clinical department. Microbiological detection of these common pathogens rarely triggers suspicion of an outbreak based on the recurrence of the pathogen. However, it must be assumed that most outbreaks occur with common pathogens. Hence, it is very likely that there is a large reporting bias favoring outbreaks with rare or MDR pathogens [21]. Highly discriminatory molecular analytical methods, e.g. whole genome sequencing, offer a possible future solution since pathogens can be quickly evaluated if suspicious clusters occur [22]. Unfortunately, for the time being these methods are time consuming, expensive, and require specialists trained in bioinformatics. Therefore, they are not currently available as routine diagnostic methods at most hospitals and often are applied only in situations where there exists high risks of further transmission or to patient health.

Outbreaks of sporadically detected pathogens

All outbreaks with sporadic pathogens were detected by 5 of 6 algorithms. NBC missed an XDR E. coli- and XDR A. baumannii outbreak. In order to provide the infection control team with a real benefit in terms of reaction time, our definition of detection required an alarm at the beginning of the outbreak. Even though the XDR A. baumannii outbreak was eventually detected, it happened only after a significant delay of more than one month (S4 Fig). The XDR E. coli outbreak was not detected by NBC at all (S2 Fig). In contrast to our results, Watkins et al. showed that NBC provides better detection rates and fewer false alarms than others, e.g. EARS [13]. Another simulation study demonstrated the superiority of NBC to EARS [23]. Those studies, however, tested NBC for natural outbreaks of disease syndromes with a high number of patients, as in natural outbreaks of viral diseases[13]. It is questionable if the circumstances of these studies are can be compared to our setting (hospital ward, bacterial pathogen, low number of cases). Further studies are needed to assess the detection detection rates of these algorithms, in particular NBC, for outbreaks in hospital wards with sporadic bacteria.

Outbreaks of endemic pathogens

Concerning outbreaks of endemic pathogens, the algorithm’s ‘normal distribution prediction interval’ was superior to all others and detected every outbreak, whereas Farrington performed worst.

Despite Farrington’s detection of 6 of 7 outbreaks, it produced too many false alarms. Our results concerning the Farrington algorithm stand in contrast to the published literature which report good performance by regression models like Farrington, e.g. in public health in France [24]. However, Farrington was designed to adjust for seasonality, which does not apply in our datasets. Moreover, regression models like Farrington do not perform well for small outbreaks (low number of patients) like those in our hospital and in most other hospital settings [25]. Previously it was shown in a simulation study that Farrington has a low sensitivity [26]. Therefore, it is possible that the Farrington algorithm alone should not be the algorithm of choice for regular hospital outbreaks.

Poisson distribution, score distribution prediction interval as well as NBC performed similarly well and showed weaknesses only regarding a VR-E. faecium outbreak. In comparison, EARS missed one additional outbreak, one of two C. difficile outbreaks. The VR-E. faecium outbreak missed occurred during an increase in VR-E. faecium incidence that was slow overall. This most likely led to an increase in false alarms in Poisson distribution and score distribution. NBC, however, reacted to this increase with fewer false alarms (higher specificity) but triggered the alarm with an unacceptable delay of more than 1.5 months after the outbreak onset.

Moreover, it should be acknowledged that in several respects, outbreaks with VRE are outstanding examples of infection control practice. First, VR-E. faecium is currently a very common pathogen with significantly increasing incidence [27] in Germany. In our datasets comprise four different outbreaks which represent the highest prevalence (up to 14 different patients per time period). Second, different VRE isolates from hospital clusters are often not distinguishable using commonly available molecular methods [28]. Moreover, because enterococci are ubiquitous bacteria, transmission routes are highly complex and include various modes of introduction between different hospitals and wards, especially oncology wards with continually returning VRE patients [28, 29]. This eventually leads to low positive predictive value for detected aberrations of clusters on oncology wards (or other high endemic settings). This is different particularly in other disciplines with low VRE prevalence. Therefore, adjusted levels of specificity based on ward discipline are required, notably for VRE in Germany, but possibly also for pathogens with similar epidemiological characteristics.

Outbreaks of Clostridium difficile infections (CDI)

Outbreaks of toxin-producing C. difficile are another problematic area of automated outbreak detection. C. difficile is a ubiquitous bacterium; the associated diarrhea (CDI) usually follows extensive antimicrobial therapy [30]. Therefore, simple pathogen detection without clinical verification of diarrhea is most likely not an adequate approach for CDI outbreak detection. Moreover, whole-genome outbreak studies showed that probably only 40% of nosocomial CDI clusters are clonal [31]. The rest are most probably independent events that follow antimicrobial therapy. They can appear as clusters due to ward-specific policies determining which antimicrobial therapies are used. Therefore, on a high endemic ward, the positive predictive value of a CDI outbreak alarm is low. Even more problematic, most laboratories examine factors other than bacterial cultures, for example toxin production. This often renders it impossible to determine post hoc clonal relatedness. We examined four different C. difficile clusters. Two of them were in sporadic situations and were detected by all algorithms (detection rate 100%). Of the two other endemic clusters, one was detected with an acceptable detection rate of 83% (outbreak 5). The other one (outbreak 4) was detected by only 50% of the algorithms we tested. A noteworthy difference between these outbreaks was the steeper increase in detection rates during outbreak 5 (compared to outbreak 4) and the associated period prior the outbreak. Our results, therefore, showed that for CDI further adjustments need to be made, in order determine the likelihood of a CDI outbreak.

Limitations

Our AODS approach is pathogen-based and can only detect possible clonal outbreaks. No gold standard for outbreak detection exists because there is no international consensus on the definition of an outbreak [3, 4, 20]. Therefore, specificity could not be calculated for our AODS. Hence, we have based our analysis on the comparison of the outbreak detection system we currently use and the AODS. Our manual outbreak detection system works prospectively but the AODS works retrospectively. This approach could have missed outbreaks. Hence, such outbreaks could be not analyzed by AODS. These Another limitation exists with regard to clonal typing which was not performed routinely. Therefore, it was not possible to detect the potential for better detection rates of the AODS with respect to our manual outbreak detection. Future prospective studies that include state-of-the-art clonal typing methods are necessary in order to assess further benefits of AODS.

Conclusion

Our work showed that an automated outbreak system for sporadic bacteria in hospitals can work reliably in many cases. It can provide an early warning system and depends only on timely reports of microbiological results. The greatest benefit of such an automated system lies in the automatic alarm for clusters of otherwise rare pathogens, especially in large hospitals. Regarding more common bacteria, the system resulted in a substantial improvement in one hospital’s outbreak detection detection rates. However, the low positive predictive value of those alarms illustrates the need for further adjustments for various other variables. For example, ward and pathogen-specific characteristics need to be taken into consideration in all analyses because they change the predictive value to a great extent. If outbreak isolates are still retrievable, the alarm would lead to further molecular analyses of the isolates’ genetic relatedness. In most cases, this will not be feasible and a decision on the likelihood of a real outbreak remains dependent on the expertise of the hospital epidemiologist. We believe that currently available technology cannot replace an experienced hospital epidemiologist. However, although it needs further development and evaluation in real life situations, our system provides fundamental work toward a system of automated outbreak detection.

Supporting information

S1 Fig [tiff]
Outbreak within a sporadic dataset (outbreak II).

S2 Fig [tiff]
Outbreak within a sporadic dataset (outbreak III).

S3 Fig [tiff]
Outbreak within a sporadic dataset (outbreak IV).

S4 Fig [tiff]
Outbreak within a sporadic dataset (outbreak V).

S5 Fig [tiff]
Outbreak within a sporadic dataset (outbreak VI).

S6 Fig [tiff]
Outbreak within a sporadic dataset (outbreak VII).

S7 Fig [tiff]
Outbreak within an endemic dataset (outbreak 1).

S8 Fig [tiff]
Outbreak within an endemic dataset (outbreak 3).

S9 Fig [tiff]
Outbreak within an endemic dataset (outbreak 4).

S10 Fig [tiff]
Outbreak within an endemic dataset (outbreak 5).

S11 Fig [tiff]
Outbreak within an endemic dataset (outbreak 6).

S12 Fig [tiff]
Outbreak within an endemic dataset (outbreak 7).

S1 Text [doc]
Detailed algorithms description.


Zdroje

1. Snitkin ES, Zelazny AM, Thomas PJ, Stock F, Henderson DK, Palmore TN, et al. Tracking a hospital outbreak of carbapenem-resistant Klebsiella pneumoniae with whole-genome sequencing. Science translational medicine. 2012;4(148):148ra16–ra16.

2. Bergin SM, Periaswamy B, Barkham T, Chua HC, Mok YM, Fung DSS, et al. An Outbreak of Streptococcus pyogenes in a Mental Health Facility: Advantage of Well-Timed Whole-Genome Sequencing Over emm Typing. Infection control and hospital epidemiology. 2018:1–9.

3. Baker MA, Huang SS, Letourneau AR, Kaganov RE, Peeples JR, Drees M, et al. Lack of comprehensive outbreak detection in hospitals. infection control & hospital epidemiology. 2016;37(4):466–8.

4. Leclere B, Buckeridge DL, Boelle PY, Astagneau P, Lepelletier D. Automated detection of hospital outbreaks: A systematic review of methods. PLoS One. 2017;12(4):e0176438. doi: 10.1371/journal.pone.0176438 28441422

5. Buckeridge DL. Outbreak detection through automated surveillance: a review of the determinants of detection. Journal of biomedical informatics. 2007;40(4):370–9. doi: 10.1016/j.jbi.2006.09.003 17095301

6. Murdoch TB, Detsky AS. The inevitable application of big data to health care. Jama. 2013;309(13):1351–2. doi: 10.1001/jama.2013.393 23549579

7. Dessau RB, Steenberg P. Computerized surveillance in clinical microbiology with time series analysis. Journal of clinical microbiology. 1993;31(4):857–60. 8463397

8. Buehler JW, Hopkins RS, Overhage JM, Sosin DM, Tong V. Framework for evaluating public health surveillance systems for early detection of outbreaks: recommendations from the CDC Working Group. MMWR Recommendations and reports: Morbidity and mortality weekly report Recommendations and reports. 2004;53(Rr-5):1–11. 15129191

9. Rutjes AW, Reitsma JB, Coomarasamy A, Khan KS, Bossuyt PM. Evaluation of diagnostic tests when there is no gold standard. A review of methods. Health technology assessment (Winchester, England). 2007;11(50):iii, ix–51.

10. Grundmann H, Barwolff S, Tami A, Behnke M, Schwab F, Geffers C, et al. How many infections are caused by patient-to-patient transmission in intensive care units? Critical care medicine. 2005;33(5):946–51. doi: 10.1097/01.ccm.0000163223.26234.56 15891318

11. Nishiura H. Early detection of nosocomial outbreaks caused by rare pathogens: a case study employing score prediction interval. Osong Public Health Res Perspect. 2012;3(3):121–7. doi: 10.1016/j.phrp.2012.07.010 24159503

12. Hutwagner L, Thompson W, Seeman GM, Treadwell T. The bioterrorism preparedness and response Early Aberration Reporting System (EARS). J Urban Health. 2003;80(2 Suppl 1):i89–96. 12791783

13. Watkins RE, Eagleson S, Veenendaal B, Wright G, Plant AJ. Applying cusum-based methods for the detection of outbreaks of Ross River virus disease in Western Australia. BMC Med Inform Decis Mak. 2008;8:37. doi: 10.1186/1472-6947-8-37 18700044

14. Pelecanos AM, Ryan PA, Gatton ML. Outbreak detection algorithms for seasonal disease data: a case study using Ross River virus disease. BMC Med Inform Decis Mak. 2010;10:74. doi: 10.1186/1472-6947-10-74 21106104

15. Farrington CP, Andrews NJ, Beale AD, Catchpole MA. A Statistical Algorithm for the Early Detection of Outbreaks of Infectious Disease. Journal of the Royal Statistical Society Series A (Statistics in Society). 1996;159(3):547–63.

16. Kulldorff M, Heffernan R, Hartman J, Assuncao R, Mostashari F. A space-time permutation scan statistic for disease outbreak detection. PLoS medicine. 2005;2(3):e59. doi: 10.1371/journal.pmed.0020059 15719066

17. Miller JK, Chen J, Sundermann A, Marsh JW, Saul MI, Shutt KA, et al. Statistical outbreak detection by joining medical records and pathogen similarity. J Biomed Inform. 2019;91:103126. doi: 10.1016/j.jbi.2019.103126 30771483

18. R Development Core Team. R: A language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing; 2019.

19. Salmon M, Schumacher D, Hohle M. Monitoring Count Time Series in R: Aberration Detection in Public Health Surveillance. J Stat Softw. 2016;70(10):1–35.

20. Stoesser N, Sheppard AE, Moore C, Golubchik T, Parry C, Nget P, et al. Extensive within-host diversity in fecally carried extended-spectrum beta-lactamase-producing Escherichia coli: implications for transmission analyses. Journal of clinical microbiology. 2015:JCM. 00378–15.

21. Gastmeier P, Stamm-Balderjahn S, Hansen S, Zuschneid I, Sohr D, Behnke M, et al. Where should one search when confronted with outbreaks of nosocomial infection? American journal of infection control. 2006;34(9):603–5. doi: 10.1016/j.ajic.2006.01.014 17097458

22. Sabat AJ, Budimir A, Nashev D, Sá-Leão R, Van Dijl JM, Laurent F, et al. Overview of molecular typing methods for outbreak detection and epidemiological surveillance. Eurosurveillance. 2013;18(4):20380. doi: 10.2807/ese.18.04.20380-en 23369389

23. Fricker RD, Hegler BL Jr., Dunfee DA. Comparing syndromic surveillance detection methods: EARS' versus a CUSUM-based methodology. Stat Med. 2008;27(17):3407–29. doi: 10.1002/sim.3197 18240128

24. Guillou A, Kratz M, Strat YL. An extreme value theory approach for the early detection of time clusters. A simulation-based assessment and an illustration to the surveillance of Salmonella. Statistics in medicine. 2014;33(28):5015–27. doi: 10.1002/sim.6275 25060768

25. Unkel S, Farrington C, Garthwaite PH, Robertson C, Andrews N. Statistical methods for the prospective detection of infectious disease outbreaks: a review. Journal of the Royal Statistical Society: Series A (Statistics in Society). 2012;175(1):49–82.

26. Bedubourg G, Le Strat Y. Evaluation and comparison of statistical methods for early temporal detection of outbreaks: A simulation-based study. PLoS One. 2017;12(7):e0181227. doi: 10.1371/journal.pone.0181227 28715489

27. Remschmidt C, Schroder C, Behnke M, Gastmeier P, Geffers C, Kramer TS. Continuous increase of vancomycin resistance in enterococci causing nosocomial infections in Germany—10 years of surveillance. Antimicrobial resistance and infection control. 2018;7:54. doi: 10.1186/s13756-018-0353-x 29760912

28. Raven KE, Gouliouris T, Brodrick H, Coll F, Brown NM, Reynolds R, et al. Complex routes of nosocomial vancomycin-resistant Enterococcus faecium transmission revealed by genome sequencing. Clinical infectious diseases. 2017;64(7):886–93. doi: 10.1093/cid/ciw872 28362945

29. Ulrich N, Vonberg RP, Gastmeier P. Outbreaks caused by vancomycin-resistant Enterococcus faecium in hematology and oncology departments: A systematic review. Heliyon. 2017;3(12):e00473. doi: 10.1016/j.heliyon.2017.e00473 29322099

30. Leffler DA, Lamont JT. Clostridium difficile infection. New England Journal of Medicine. 2015;372(16):1539–48. doi: 10.1056/NEJMra1403772 25875259

31. Eyre DW, Cule ML, Walker AS, Crook DW, Wilcox MH, Peto TE. Hospital and community transmission of Clostridium difficile: a whole genome sequencing study. The Lancet. 2012;380:S12.


Článek vyšel v časopise

PLOS One


2020 Číslo 1
Nejčtenější tento týden
Nejčtenější v tomto čísle
Kurzy

Zvyšte si kvalifikaci online z pohodlí domova

Důležitost adherence při depresivním onemocnění
nový kurz
Autoři: MUDr. Eliška Bartečková, Ph.D.

Svět praktické medicíny 1/2024 (znalostní test z časopisu)

Koncepce osteologické péče pro gynekology a praktické lékaře
Autoři: MUDr. František Šenk

Sekvenční léčba schizofrenie
Autoři: MUDr. Jana Hořínková

Hypertenze a hypercholesterolémie – synergický efekt léčby
Autoři: prof. MUDr. Hana Rosolová, DrSc.

Všechny kurzy
Kurzy Podcasty Doporučená témata Časopisy
Přihlášení
Zapomenuté heslo

Zadejte e-mailovou adresu, se kterou jste vytvářel(a) účet, budou Vám na ni zaslány informace k nastavení nového hesla.

Přihlášení

Nemáte účet?  Registrujte se

#ADS_BOTTOM_SCRIPTS#