Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Monday, December 24 through Wednesday, December 26 inclusive. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Advertisement

Citing this Article

Right click to copy or hit: ctrl+c (cmd+c on mac)

Published on 21.09.15 in Vol 3, No 3 (2015): Jul-Sep

This paper is in the following e-collection/theme issue:

    Original Paper

    A Web-Based, Hospital-Wide Health Care-Associated Bloodstream Infection Surveillance and Classification System: Development and Evaluation

    1Graduate Institute of Biomedical Electronics and Bioinformatics, National Taiwan University, Taipei, Taiwan

    2Computational Health Informatics Program, Boston Children’s Hospital, Boston, MA, United States

    3Department of Electrical Engineering, National Taiwan University, Taipei, Taiwan

    4Center for Infection Control, National Taiwan University Hospital, Taipei, Taiwan

    5Information Systems Office, National Taiwan University Hospital, Taipei, Taiwan

    6Department of Computer Science and Information Engineering, National Taiwan University, Taipei, Taiwan

    7Department of Internal Medicine, National Taiwan University Hospital and College of Medicine, Taipei, Taiwan

    8National Institute of Infectious Diseases and Vaccinology, National Health Research Institutes, Miaoli, Taiwan

    Corresponding Author:

    Yee-Chun Chen, MD, PhD

    Department of Internal Medicine

    National Taiwan University Hospital and College of Medicine

    No. 7 Chung-Shan South Road

    Taipei, 100

    Taiwan

    Phone: 886 223123456 ext 65054

    Fax:886 223971412

    Email:


    ABSTRACT

    Background: Surveillance of health care-associated infections is an essential component of infection prevention programs, but conventional systems are labor intensive and performance dependent.

    Objective: To develop an automatic surveillance and classification system for health care-associated bloodstream infection (HABSI), and to evaluate its performance by comparing it with a conventional infection control personnel (ICP)-based surveillance system.

    Methods: We developed a Web-based system that was integrated into the medical information system of a 2200-bed teaching hospital in Taiwan. The system automatically detects and classifies HABSIs.

    Results: In this study, the number of computer-detected HABSIs correlated closely with the number of HABSIs detected by ICP by department (n=20; r=.999 P<.001) and by time (n=14; r=.941; P<.001). Compared with reference standards, this system performed excellently with regard to sensitivity (98.16%), specificity (99.96%), positive predictive value (95.81%), and negative predictive value (99.98%). The system enabled decreasing the delay in confirmation of HABSI cases, on average, by 29 days.

    Conclusions: This system provides reliable and objective HABSI data for quality indicators, improving the delay caused by a conventional surveillance system.

    JMIR Med Inform 2015;3(3):e31

    doi:10.2196/medinform.4171

    KEYWORDS



    Introduction

    Background

    Health care-associated infections (HAIs), adverse events related to health care, excess mortality and morbidity, and resource use are responsible for augmenting antimicrobial resistance [1,2]. Surveillance of HAIs is an essential component of infection control programs in health care settings. The goals of surveillance are to assess the disease incidence, identify the niche and opportunity for improvement, monitor the efficacy of interventions, and support the rationale behind changes in policies [3]. Previous studies have reported that HAIs have decreased ly in hospitals that adopted surveillance programs in the 1980s in the United States [4,5]. In 1981, a hospital-wide HAI surveillance program was initiated at the National Taiwan University Hospital (NTUH). Factors such as periodic feedback to the departments or wards and intensified interventions resulted in a decrease in surgical-site and respiratory tract infections in the 1980s [6]. Following the upgrading of the infection prevention and control program in 2004, there was a significant reduction in bloodstream infections, HAIs in intensive care units, and HAIs caused by multidrug-resistant organisms (MDROs) during the period from 2004 to 2007 [7].

    Attention to HAIs has increased partially because of legislative mandates for reporting and reimbursement policies [8]. However, conventional HAI surveillance systems require considerable human involvement in integrating and interpreting data and are labor intensive, performance dependent, and tend to divert resources that are necessary for implementing control measures and prevention activities [9]. Relying on employees in institutions in an environment where reporting HAIs can be associated with punitive consequences is suboptimal [10]. Furthermore, the decision rules (ie, the case definitions) of HAIs are complicated when the complexity of the current health care in tertiary care hospitals is considered. Utilization of hospital discharge registry data delays the detection of HAI, eventually resulting in ineffective identification of problems [11]. Recent studies have identified interinstitutional variability in surveillance techniques, an inconsistency that affects the validity of publicly reported HAI data [12]. Developing reliable and objective HAI definitions and automated processes for infection detection is crucial; however, transformation into automated surveillance system remains challenging [8,10].

    Study Objective

    In continuation with our previous efforts in developing a Web-based MDRO surveillance system that automatically identifies and accurately detects suspicious outbreaks of MDROs [13], implementation of which could save 1 person-day daily, we conducted this study with the aim of developing a Web-based automatic surveillance and classification system for health care-associated bloodstream infection (HABSI), the most common HAIs at NTUH [6]. In addition, performance of the system was evaluated by comparing the proposed system with a conventional infection control personnel (ICP)-based surveillance system.


    Methods

    Hospital Setting and Study Population

    The study was conducted at NTUH, a 2200-bed teaching hospital that provides primary and tertiary care for the adult and the pediatric population in Taiwan. This study was approved by the Institutional Ethics Review Board of NTUH (NTUH-200904014R). In 2011, NTUH served 2,309,108 outpatients, received 106,950 emergency visits, and discharged 104,899 patients (723,505 patient-days).

    Two sets of blood samples from separate venipuncture sites for bacterial culture were routinely collected from patients who were newly diagnosed with sepsis. An additional sample was collected after 45-60 minutes to define continuous bloodstream infection (BSI). Only 1 blood sample for a follow-up culture was collected to confirm the clearance of BSI. Of the 80,327 blood specimens that were sent for isolation and identification of pathogens, 991 (1.23%) were obtained through a single blood draw. A total of 8745 samples grew 1 or more pathogens (10.88%); of these, 1908 exhibited HABSIs. The pooled mean of HABSI incidence was 14.7 episodes/1000 patients (range 0.2-112.7/1000 patients by department) and 2.13 episodes/1000 patient-days.

    Conventional ICP-Based HAI Surveillance System

    A prospective, hospital-wide on-site surveillance of HAIs, initiated in 1981 [6], was conducted by ICPs who reviewed microbiological data daily and visited inpatient units weekly to identify patients with HAIs according to definitions of the Centers for Disease Control and Prevention (CDC) [14] before we implemented the Web-based surveillance system. If required, the ICPs consulted physicians, particularly infectious disease physicians, to interpret the medical information of patients who have received complicated interventions. Data are collected on standardized data-collection forms and input into the computer database manually. In addition, the ICPs monitor culture results from the clinical microbiology laboratory daily to identify MDROs. The unit-specific incidences of HAIs, including overall, unit-based, and site-specific infection rates, are analyzed monthly and compared with historical data. Feedback is provided to each unit to stimulate intervention measures.

    The key data are collected by systematically reviewing hospital information systems (HISs), laboratory information systems (LISs), and handwritten charts. Because of advances in medical information system, data of HISs and LISs are currently stored as electronic medical records (EMRs). However, data generated in the medical information system are scattered in numerous databases, and data access is hindered by several interfaces. In addition, data must be integrated, interpreted, and transformed into meaningful information.

    Web-Based HAI Surveillance and Classification System

    We established a rule-based HABSI surveillance and classification system (the system), which was implemented on October 1, 2010. The current version was revised on September 20, 2012 (Figure 1). The system screens HIS and LIS data daily to detect HABSI candidates according to the well-defined detection rules. The system detects and classifies HABSIs automatically and reserves professional autonomy by requiring further confirmation. Figure 1 shows the general architecture of the system, including the user layer, the database layer, and the 3-part system layer (data collection, candidate detection, and HAI management). The system adheres to service-oriented architecture (SOA) and Health Level Seven (HL7) standards and can be adapted in other information systems [15].

    This data-collection subsystem collects relevant data from HIS and LIS using HL7 standards, which was extended stepwise from the previous version [13]. For example, data related to age, sex, ward transfer, admission date, and discharge diagnosis are obtained from the hospital administrative system. Re-admissions within 2 days of discharge are linked to the previous hospital stay and considered to be a single hospitalization episode. Procedure codes are obtained from the hospital billing system. Data on body temperature, heart rate, respiratory rate, and presence of device including indwelling catheters are obtained from the nursing system. The use of antimicrobial agents is obtained from the pharmacology prescription system. The laboratory data consist of specimen information and microbiological data. We used the specimen log-in time as a proxy for the time of specimen collection at bedside and the infection time.

    To develop the detection rules in the candidate-detection subsystem, the ICPs have adapted objective components of the National Health Care Safety Network (NHSN) definitions established by the CDC [14] and modified them according to the consensus between Taiwan Center for Disease Control and NTUH (see Multimedia Appendix 1 for detection rules and Multimedia Appendices 2 and 3 for list of devices and signs and symptoms, respectively). Computer engineers have established HABSI detection rules accordingly (Figure 2). Thus, there are differences between the detection rules and NHSN definition. Although the secondary HABSI was removed from the NHSN definition in 2008, we used the data in the NTUH infection control system to maintain data consistency for time trend analysis [6,7]. The primary HABSI was divided into 3 subtypes for quality-improvement purpose, including central line-related BSI (CRBSI), central line-associated BSI (other than CRBSI), and other primary HABSI. Detection rules did not include free text data that were not available in the EMRs, such as chills and apnea, during the study period.

    The HAI management subsystem, established on July 1, 2007 [13], and upgraded periodically (refer to the screenshot in Multimedia Appendix 4), consists of data integration, display, and a data-modification user interface for facilitating the surveillance of HAIs. The HAI management subsystem has a single entry point for HIS through the browser [16] and comprises a storage information component relevant to HAI. The information for each event of HAI includes patient demographic data, diagnoses, procedures, medications, and microbiology reports to facilitate the confirmation of HAI by ICPs. If required, the ICPs can modify these HABSI cases or add additional HABSIs that were undetectable by the subsystem. The latter might occur for new units, new pathogens, new procedures, or other elements, which are not yet included in the current database. Furthermore, the system provides data analysis and process-control charts [13].

    Figure 1. General architecture of the Web-based health care-associated infection (HAI) surveillance and classification system. DB: database; HL7: Health Level Seven; SOA: service-oriented architecture.
    View this figure
    Figure 2. Computer algorithms to detect health care-associated bloodstream infection (HABSI) by active daily screening of data from hospital information system and laboratory information system. HABSIs are classified into primary HABSI (PRIM), secondary HABSI (SEC), and clinical sepsis (CSEP) as described in Multimedia Appendix 1. Polymicrobial and persistent BSI criterion here are to eliminate false signals due to duplicate counting, etc.
    View this figure

    Evaluation of System Performance and Statistical Analyses

    Clinically useful tests must be valid and reliable and have a reasonable turnaround time. Thus, we conducted a 3-aspect evaluation, including accuracy, reliability, and efficiency, of the system. Figure 3 summarizes the objectives, methods, and evaluation periods. Reliability of the system was evaluated before implementation of the first version of the system in 2010; accuracy of HABSI rules was evaluated in October 2012. Furthermore, we evaluated, and continue to evaluate, the stepwise improvement in efficiency after implementation of the HAI management system and the HABSI surveillance and classification system.

    We first evaluated the performance of the system during the developmental phase (ie, before implementation of the system) regarding its potential to provide data for quality indicators. Computer-detected HABSIs were compared with ICP-detected HABSIs as a routine practice between July 1, 2010, and September 30, 2010. The correlation between these 2 data sources was analyzed according to department distribution and time trend of HABSIs.

    On the basis of inconsistent and varied performances of the conventional ICP-based surveillance system, we further evaluated the performance after implementing the system using ICP-defined reference standards. To generate high-quality reference standards, 11 ICPs performed a retrospective review of all medical data of patients who were admitted between October 1, 2012, and October 31, 2012, to identify HABSI cases based on NTUH detection rules, and one of the authors (H-CL) validated the results. The sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV) of the system, and Cohen kappa coefficient [17] were calculated in this evaluation method. All the performance indicators were calculated based on whether a patient had HABSI or not.

    We then compared the delay in HABSI confirmation (as a proxy for practice efficiency) before (October 2007-September 2010) and after (October 2010-September 2013) the system implementation. The delay in confirmation was defined as the intervals between the HABSI confirmation dates (complete data entry and confirmed by ICPs in the HAI management subsystem) and log-in dates of the first blood specimen with positive results and was calculated by averaging the number of delay days in each month. The study periods were selected taking into account the seasonal variation of HAI rates demonstrated previously [7].

    Figure 3. Timeline of development and performance evaluation of the health care-associated bloodstream infections (HABSIs) surveillance and classification system. ICP: infection control personnel.
    View this figure

    Results

    Performance of the System as a Provider of Quality Indicator

    During the 14-week study period (July 1, 2010-September 30, 2010), 501 episodes of ICP-detected HABSIs and 479 episodes of computer-detected HABSIs were found throughout the 20 departments. The data were highly correlated by place and time (Figure 4). These results indicated that the data provided by the system are suitable quality indicators. Thus, we implemented the system on October 1, 2010.

    Figure 4. Correlation of 501 episodes of infection control personnel (ICP)-detected health care-associated bloodstream infection (HABSI) and 479 episodes of computer-detected HABSI from 20 departments during the 14-week study period. (A) Perfect agreement of HABSI episodes by department (n=20, Pearson correlation, r<.999, P<.001). (B) Perfect agreement of HABSI episodes by time (n=14, Pearson correlation, r=.941, P<.001).
    View this figure

    Accuracy of the Detection Rule

    In October 2012, the system identified 167 episodes of HABSIs (Figure 5), including 160 of 163 reference standard episodes in 31 days (Table 1). The sensitivity and specificity of the HABSI classification system were 98.16% (95% CI 94.29-99.52) and 99.96% (95% CI 99.91-99.98), respectively. The PPV and NPV were 95.81% (95% CI 91.22-98.15) and 99.98% (95% CI 99.95-100.00), respectively. Moreover, the agreement between the computer-detected HABSIs and the reference standard was nearly perfect (Cohen kappa coefficient .97; 95% CI 0.95-0.99). The performance of the system for detecting central line-associated HABSI was also excellent (sensitivity 97.14%, specificity 99.94%, PPV 91.07%, NPV 99.94%, and Cohen kappa coefficient .94).

    Table 1. Comparison of the case detection results of the health care-associated bloodstream infection surveillance and classification system with infection control personnel reference standard between the periods October 1 and October 31, 2012.
    View this table
    Figure 5. Computer algorithms identifying 167 events of health care-associated bloodstream infection among 8862 inpatients and 9132 patients in the emergency department between October 1 and October 31, 2012 (31 days).
    View this figure

    Decrease in the Delay of HABSI Confirmation

    The delay in HABSI confirmation was reduced from 43.58 (SD 15.57) days before the system implementation (October 2007-September 2010, 1096 days) to 14.58 (SD 4.64) days after the implementation (October 2010-September 2013, 1096 days; P<.001). Figure 6 shows that the time trend of the delay in HABSI confirmation, which was as high as 90 days in July 2007, decreased after the implementation of the HAI management subsystem in July 2007, and further improved after automating the system in October 2010. The delay in the second half of 2013 was only 5.78 (SD 0.91) days. Conversely, without the system, the delay increased during H1N1 influenza pandemics and when preparing for international accreditation.

    Figure 6. The detection delay of health care-associated bloodstream infection (HABSI) decreased gradually from July 2007 to December 2013. The first version of health care-associated infection (HAI) management subsystem has been developed to facilitate infection control personnel-based surveillance program since July 2007. This was revised stepwise and has been operation as an automatic system since October, 2010. In March 2009, this hospital initiated preparedness for international hospital accreditation, which was scheduled 1 year later. Influenza pandemic occurred in April 2009.
    View this figure

    Discussion

    Preliminary Findings

    This Web-based, automated HABSI surveillance and classification system used discrete data elements obtained from HISs, and LISs provided data highly correlated with conventional ICP surveillance system. The performance was excellent regarding sensitivity, specificity, PPV, and NPV, and was in agreement with reference standards; the system reduced the delay in confirmation, on average, by 29 days. The system improves practice efficiency, enabling ICPs to intensify intervention and further reduce HAI rates.

    Computer-assisted HAI surveillance and classification systems are widely implemented [11,18-35]. Studies have demonstrated using various algorithms for detecting HAIs, although with varied outcomes (summary in Multimedia Appendix 5) [25-35]. The most critical performance characteristics of these kinds of surveillance systems are sensitivity and NPV; the efficiency of the system can be assessed according to the PPV [36]. Compared with reference standards, the current version of detection rules and computer algorithms performed excellently with regard to these 3 parameters (sensitivity 98.16%, NPV 99.98%, and PPV 95.81%) because of the following reasons: Through cross-talk among ICPs, infectious disease physicians, and engineers, we integrated clinical know-how and translated international case definitions to define detection rules and construct computer algorithms. The system was established and revised through a plan-do-check-act cycle in a general hospital, which serves a varied patient population and offers numerous procedures. Furthermore, we evaluated the clinical utility of the system, comparing it with the prospective, hospital-wide, conventional surveillance system and reference standards.

    In this study, the HABSI detection rules (see Multimedia Appendix 1) were clearly defined, and computer algorithms (Figure 2) provided excellent results. We adapted the US CDC definition of HAI, adding rules related to re-admission within 48 hours and neonates, and a “hospital-acquired” rule, defined as a positive blood culture that was obtained 48 hours or more after admission. These rules included rules for clinical sepsis, and the system actively screened the heart rate, respiratory rate, and body temperature for infection-related symptoms and signs. These efforts facilitated ameliorating the potential underestimation of HAI when only laboratory data were used. The HABSI detection rules included polymicrobial and persistent BSI criterion to eliminate the majority of false signals (eg, duplicate counting). In addition, classification of HABSIs is flexible to addressing local policy and ICP requests to compare them with the NTUH historical data.

    The system detects and classifies HABSIs automatically and ensures professional autonomy by requiring further confirmation. Each episode of HAI requires confirmation by ICPs. The system presents detailed information about each HABSI candidate systematically to support decision making. The main reason for this design is because the system is imperfect (see the “Limitations” section). Furthermore, because HAIs are rare in hospitalized patients, the system aims to select potential HAI candidates and exclude patients who do not have an HAI and hence do not require review by ICPs.

    This study verifies the potential of the system to provide data for quality indicators. The system enabled sustainable surveillance, generating data that were correlated with conventional surveillance systems by department and time. In addition, the delay in HABSI confirmation decreased to 5.78 (SD 0.91) days in the second half of 2013. Because of the reduced length of hospital stay and the increased threat of emerging infectious diseases, early detection of HAIs can enable identifying the reservoir or index case and providing early intervention before pathogens spread further. Currently, the delay in HABSI confirmation is caused by the time required to identify the positive blood cultures and microorganisms; the system detects and analyzes results of blood cultures to prevent false alerts. Furthermore, the major challenge encountered when sharing automated HABSI surveillance systems between hospitals is different HIS settings [22]; the Web-based system, which adheres to SOA and HL7 standards, can be easily extended to and adapted for use with other medical information systems.

    Limitations

    Although our results suggested that the system performs well, this study had several limitations. First, data integrity and instantaneity substantially affect performance, because this system uses EMRs from many sources. In addition, not all data required for HAI surveillance [37] are available on the Web, because EMRs were not fully implemented at NTUH until January 1, 2014. Second, the quality of source data, which is related to the performance of clinical practice and EMRs, affected the results. As much as 1.2% (991/80,327) of the blood specimens collected in 2011 were obtained through a single blood draw (reasons described in the “Methods” section), and this affected the identification of common skin commensals and resulted in a false-positive HABSI case (Table 1). We reviewed the medical records and determined that the false-positive result was caused by the delay in specimen delivery and log-in time. Third, the system updates laboratory data from LIS within a fixed period (recall day); however, data are not repeated during a subsequent hospital stay. This resulted in 4 false-positive HABSI cases, because the laboratory reports were revised in the LIS after recall day and included common skin commensal contaminants that were not updated in the system.

    Fourth, the agreement regarding the place of onset (responsible ward) was not evaluated in this study, because patients are frequently transferred to different wards and electronic clinical data regarding symptoms and signs of infection were unavailable during the study period. Fifth, the case definition of HABSI is complicated when clinical scenarios are taken into consideration. For disagreements and received revaluations (Table 1), all the false-negative results were due to the case definition of fever (>38°C). Furthermore, we did not evaluate the reduction of person-hours after implementing the system, as described in previous studies [13,38,39], or the subsequent effect on the reduction of HABSI by reallocating ICP time and effort from collecting data to improving program quality, as described in a previous research of a hospital-wide hand-hygiene promotion program [7] and care bundles for device-associated infection to prevent HABSI.

    Conclusions

    This fully automated system that can be integrated in medical information systems detects and classifies HABSI within 5.78 (SD 0.91) days after occurrence, enabling the opportunity for early intervention. Currently, the system and other components of the infection control system [13,39] operate well and have become indispensable tools for infection control programs. Future studies using clinical data from complete EMRs and refining classification algorithms or adopting multivariable prediction models are warranted [36]. According to the results of this pilot study for HABSI automated surveillance, further efforts for other HAI surveillance are underway.

    Acknowledgments

    The authors are grateful to the members of the Center of Infection Control and the Information Systems Office at National Taiwan University Hospital for their contributions to this study, and to Kenneth D Mandl for his critical review and comments. This work was supported by grants from the Department of Health, Taiwan (Grant Nos DOH98-DC-1007 and DOH100-TD-B-111-001 to Y-CC) and the National Science Council, Taiwan (Grant Nos NSC 101-2221-E-002 -203 -MY3 to FL and NSC 103-2917-I-564-063 to Y-JT). The funding organizations were not involved in designing or conducting the study, data collection, management, analysis, or interpretation, or in the preparation, review, and approval of the manuscript.

    Conflicts of Interest

    None declared.

    Multimedia Appendix 1

    Detection rules for the health care-associated bloodstream infection (HABSI) Surveillance and Classification System at National Taiwan University Hospital (NTUH) and corresponding Centers for Disease Control and Prevention (CDC) National Healthcare Safety Network (NHSN) definition of health care-associated infection (HAI) [14].

    PDF File (Adobe PDF File), 53KB

    Multimedia Appendix 2

    The list of central lines and intravascular devices.

    PDF File (Adobe PDF File), 29KB

    Multimedia Appendix 3

    The definition of symptoms/signs of sepsis in pediatric patients.

    PDF File (Adobe PDF File), 20KB

    Multimedia Appendix 4

    Summary of characteristics and performances of computer-assisted health care-associated infection surveillance systems in the literature.

    PDF File (Adobe PDF File), 40KB

    Multimedia Appendix 5

    User interface of HAI Management System. This system integrated all the information which was needed for HAI decision.

    PDF File (Adobe PDF File), 210KB

    References

    1. European Centre for Disease Prevention and Control. European Centre for Disease Prevention and Control: Annual Epidemiological Report on Communicable Diseases in Europe 2008. Stockholm, Sweden; 2008.   URL: http://ecdc.europa.eu/en/publications/Publications/0812_SUR_Annual_Epidemiological_Report_2008.pdf [accessed 2014-12-23] [WebCite Cache]
    2. Klevens RM, Edwards JR, Richards CL, Horan TC, Gaynes RP, Pollock DA, et al. Estimating health care-associated infections and deaths in U.S. hospitals, 2002. Public Health Rep 2007;122(2):160-166 [FREE Full text] [Medline]
    3. McKibben L, Horan TC, Tokars JI, Fowler G, Cardo DM, Pearson ML, Healthcare Infection Control Practices Advisory Committee. Guidance on public reporting of healthcare-associated infections: Recommendations of the Healthcare Infection Control Practices Advisory Committee. Infect Control Hosp Epidemiol 2005 Jun;26(6):580-587. [CrossRef] [Medline]
    4. Gaynes R, Richards C, Edwards J, Emori TG, Horan T, Alonso-Echanove J, et al. Feeding back surveillance data to prevent hospital-acquired infections. Emerg Infect Dis 2001;7(2):295-298 [FREE Full text] [CrossRef] [Medline]
    5. Haley RW, Culver DH, White JW, Morgan WM, Emori TG, Munn VP, et al. The efficacy of infection surveillance and control programs in preventing nosocomial infections in US hospitals. Am J Epidemiol 1985 Feb;121(2):182-205. [Medline]
    6. Chuang YC, Chen YC, Chang SC, Sun CC, Chang YY, Chen ML, et al. Secular trends of healthcare-associated infections at a teaching hospital in Taiwan, 1981-2007. J Hosp Infect 2010 Oct;76(2):143-149. [CrossRef] [Medline]
    7. Chen YC, Sheng WH, Wang JT, Chang SC, Lin HC, Tien KL, et al. Effectiveness and limitations of hand hygiene promotion on decreasing healthcare-associated infections. PLoS One 2011;6(11):e27163 [FREE Full text] [CrossRef] [Medline]
    8. Talbot TR, Bratzler DW, Carrico RM, Diekema DJ, Hayden MK, Huang SS, Healthcare Infection Control Practices Advisory Committee. Public reporting of health care-associated surveillance data: Recommendations from the healthcare infection control practices advisory committee. Ann Intern Med 2013 Nov 5;159(9):631-635. [CrossRef] [Medline]
    9. Emori TG, Edwards JR, Culver DH, Sartor C, Stroud LA, Gaunt EE, et al. Accuracy of reporting nosocomial infections in intensive-care-unit patients to the National Nosocomial Infections Surveillance System: A pilot study. Infect Control Hosp Epidemiol 1998 May;19(5):308-316. [Medline]
    10. Trick WE. Decision making during healthcare-associated infection surveillance: A rationale for automation. Clin Infect Dis 2013 Aug;57(3):434-440 [FREE Full text] [CrossRef] [Medline]
    11. Goto M, Ohl ME, Schweizer ML, Perencevich EN. Accuracy of administrative code data for the surveillance of healthcare-associated infections: A systematic review and meta-analysis. Clin Infect Dis 2014 Mar;58(5):688-696 [FREE Full text] [CrossRef] [Medline]
    12. Hebden JN. Rationale for accuracy and consistency in applying standardized definitions for surveillance of health care-associated infections. Am J Infect Control 2012 Jun;40(5 Suppl):S29-S31. [CrossRef] [Medline]
    13. Tseng YJ, Wu JH, Ping XO, Lin HC, Chen YY, Shang RJ, et al. A Web-based multidrug-resistant organisms surveillance and outbreak detection system with rule-based classification and clustering. J Med Internet Res 2012;14(5):e131 [FREE Full text] [CrossRef] [Medline]
    14. Horan TC, Andrus M, Dudeck MA. CDC/NHSN surveillance definition of health care-associated infection and criteria for specific types of infections in the acute care setting. Am J Infect Control 2008 Jun;36(5):309-332. [CrossRef] [Medline]
    15. Health Level Seven. HL7 Standard v2.5. 2003.   URL: http://www.hl7.org/ [accessed 2014-12-23] [WebCite Cache]
    16. Hsieh SH, Hsieh SL, Cheng PH, Lai F. E-health and healthcare enterprise information system leveraging service-oriented architecture. Telemed J E Health 2012 Apr;18(3):205-212. [CrossRef] [Medline]
    17. Carletta J. Assessing agreement on classification tasks: The kappa statistic. Comput Linguist 1996 Jun 1;22(2):249-254.
    18. Steurbaut K, Van HS, Colpaert K, Lamont K, Taveirne K, Depuydt P, et al. Use of web services for computerized medical decision support, including infection control and antibiotic management, in the intensive care unit. J Telemed Telecare 2010;16(1):25-29. [CrossRef] [Medline]
    19. Bellini C, Petignat C, Francioli P, Wenger A, Bille J, Klopotov A, et al. Comparison of automated strategies for surveillance of nosocomial bacteremia. Infect Control Hosp Epidemiol 2007 Sep;28(9):1030-1035. [CrossRef] [Medline]
    20. Chalfine A, Cauet D, Lin WC, Gonot J, Calvo-Verjat N, Dazza F, et al. Highly sensitive and efficient computer-assisted system for routine surveillance for surgical site infection. Infect Control Hosp Epidemiol 2006 Aug;27(8):794-801. [CrossRef] [Medline]
    21. Apte M, Landers T, Furuya Y, Hyman S, Larson E. Comparison of two computer algorithms to identify surgical site infections. Surg Infect (Larchmt) 2011 Dec;12(6):459-464 [FREE Full text] [CrossRef] [Medline]
    22. Hota B, Lin M, Doherty JA, Borlawsky T, Woeltje K, Stevenson K, CDC Prevention Epicenter Program. Formulation of a model for automating infection surveillance: Algorithmic detection of central-line associated bloodstream infection. J Am Med Inform Assoc 2010;17(1):42-48 [FREE Full text] [CrossRef] [Medline]
    23. Klompas M, Yokoe DS. Automated surveillance of health care-associated infections. Clin Infect Dis 2009 May 1;48(9):1268-1275 [FREE Full text] [CrossRef] [Medline]
    24. Brossette SE, Hacek DM, Gavin PJ, Kamdar MA, Gadbois KD, Fisher AG, et al. A laboratory-based, hospital-wide, electronic marker for nosocomial infection: The future of infection control surveillance? Am J Clin Pathol 2006 Jan;125(1):34-39 [FREE Full text] [Medline]
    25. Woeltje KF, McMullen KM, Butler AM, Goris AJ, Doherty JA. Electronic surveillance for healthcare-associated central line-associated bloodstream infections outside the intensive care unit. Infect Control Hosp Epidemiol 2011 Nov;32(11):1086-1090. [CrossRef] [Medline]
    26. Choudhuri JA, Pergamit RF, Chan JD, Schreuder AB, McNamara E, Lynch JB, et al. An electronic catheter-associated urinary tract infection surveillance tool. Infect Control Hosp Epidemiol 2011 Aug;32(8):757-762. [CrossRef] [Medline]
    27. McBryde ES, Brett J, Russo PL, Worth LJ, Bull AL, Richards MJ. Validation of statewide surveillance system data on central line-associated bloodstream infection in intensive care units in Australia. Infect Control Hosp Epidemiol 2009 Nov;30(11):1045-1049. [CrossRef] [Medline]
    28. Pokorny L, Rovira A, Martín-Baranera M, Gimeno C, Alonso-Tarrés C, Vilarasau J. Automatic detection of patients with nosocomial infection by a computer-based surveillance system: A validation study in a general hospital. Infect Control Hosp Epidemiol 2006 May;27(5):500-503. [CrossRef] [Medline]
    29. Leth RA, Møller JK. Surveillance of hospital-acquired infections based on electronic hospital registries. J Hosp Infect 2006 Jan;62(1):71-79. [CrossRef] [Medline]
    30. Trick WE, Zagorski BM, Tokars JI, Vernon MO, Welbel SF, Wisniewski MF, et al. Computer algorithms to detect bloodstream infections. Emerg Infect Dis 2004 Sep;10(9):1612-1620 [FREE Full text] [CrossRef] [Medline]
    31. de Bruin JS, Adlassnig K, Blacky A, Mandl H, Fehre K, Koller W. Effectiveness of an automated surveillance system for intensive care unit-acquired infections. J Am Med Inform Assoc 2013;20(2):369-372 [FREE Full text] [CrossRef] [Medline]
    32. Bouam S, Girou E, Brun-Buisson C, Karadimas H, Lepage E. An intranet-based automated system for the surveillance of nosocomial infections: Prospective validation compared with physicians' self-reports. Infect Control Hosp Epidemiol 2003 Jan;24(1):51-55. [CrossRef] [Medline]
    33. van Mourik MS, Moons KG, van Solinge WW, Berkelbach-van der Sprenkel JW, Regli L, Troelstra A, et al. Automated detection of healthcare associated infections: External validation and updating of a model for surveillance of drain-related meningitis. PLoS One 2012;7(12):e51509 [FREE Full text] [CrossRef] [Medline]
    34. Du M, Xing Y, Suo J, Liu B, Jia N, Huo R, et al. Real-time automatic hospital-wide surveillance of nosocomial infections and outbreaks in a large Chinese tertiary hospital. BMC Med Inform Decis Mak 2014;14:9 [FREE Full text] [CrossRef] [Medline]
    35. Woeltje KF, Butler AM, Goris AJ, Tutlam NT, Doherty JA, Westover MB, et al. Automated surveillance for central line-associated bloodstream infection in intensive care units. Infect Control Hosp Epidemiol 2008 Sep;29(9):842-846. [CrossRef] [Medline]
    36. van Mourik MS, Troelstra A, van Solinge WW, Moons KG, Bonten MJ. Automated surveillance for healthcare-associated infections: Opportunities for improvement. Clin Infect Dis 2013 Jul;57(1):85-93 [FREE Full text] [CrossRef] [Medline]
    37. Woeltje KF, Lin MY, Klompas M, Wright MO, Zuccotti G, Trick WE. Data requirements for electronic surveillance of healthcare-associated infections. Infect Control Hosp Epidemiol 2014 Sep;35(9):1083-1091. [CrossRef] [Medline]
    38. Hajdu A, Eriksen HM, Sorknes NK, Hauge SH, Loewer HL, Iversen BG, et al. Evaluation of the national surveillance system for point-prevalence of healthcare-associated infections in hospitals and in long-term care facilities for elderly in Norway, 2002-2008. BMC Public Health 2011;11:923 [FREE Full text] [CrossRef] [Medline]
    39. Tseng YJ, Chen YC, Lin HC, Wu JH, Chen MY, Lai F. A web-based hospital-acquired infection surveillance information system. New York: IEEE; 2010 Oct Presented at: 10th IEEE International Conference on Information Technology and Applications in Biomedicine (ITAB); Nov 3-5, 2010; Corfu, Greece p. 1-4. [CrossRef]


    Abbreviations

    CABSI: central line-associated bloodstream infection
    CDC: Disease Control and Prevention
    CRBSI: central line-related bloodstream infection
    EMRs: electronic medical records
    HABSI: health care-associated bloodstream infection
    HAIs: health care-associated infections
    HIS: hospital information systems
    HL7: Health Level Seven
    ICP: infection control personnel
    LISs: laboratory information systems
    MDROs: multidrug-resistant organism
    NHSN: National Health Care Safety Network
    NPV: negative predictive value
    NTUH: National Taiwan University Hospital
    PPV: positive predictive value
    SOA: service-oriented architecture


    Edited by G Eysenbach; submitted 23.12.14; peer-reviewed by W Trick, J de Bruin; comments to author 07.04.15; revised version received 07.06.15; accepted 24.07.15; published 21.09.15

    ©Yi-Ju Tseng, Jung-Hsuan Wu, Hui-Chi Lin, Ming-Yuan Chen, Xiao-Ou Ping, Chun-Chuan Sun, Rung-Ji Shang, Wang-Huei Sheng, Yee-Chun Chen, Feipei Lai, Shan-Chwen Chang. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 21.09.2015.

    This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on http://medinform.jmir.org/, as well as this copyright and license information must be included.