Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2017 Apr:68:112-120.
doi: 10.1016/j.jbi.2017.03.009. Epub 2017 Mar 16.

Strategies for handling missing clinical data for automated surgical site infection detection from the electronic health record

Affiliations

Strategies for handling missing clinical data for automated surgical site infection detection from the electronic health record

Zhen Hu et al. J Biomed Inform. 2017 Apr.

Abstract

Proper handling of missing data is important for many secondary uses of electronic health record (EHR) data. Data imputation methods can be used to handle missing data, but their use for analyzing EHR data is limited and specific efficacy for postoperative complication detection is unclear. Several data imputation methods were used to develop data models for automated detection of three types (i.e., superficial, deep, and organ space) of surgical site infection (SSI) and overall SSI using American College of Surgeons National Surgical Quality Improvement Project (NSQIP) Registry 30-day SSI occurrence data as a reference standard. Overall, models with missing data imputation almost always outperformed reference models without imputation that included only cases with complete data for detection of SSI overall achieving very good average area under the curve values. Missing data imputation appears to be an effective means for improving postoperative SSI detection using EHR clinical data.

Keywords: Electronic health records; Missing data; Surgical site infections.

PubMed Disclaimer

Figures

Figure 1
Figure 1
Detection Performance for each category of SSI with different imputation methods. The AUC scores are calculated based on both the training set (using the 10-fold cross validation) and the test set. Generally, the results indicate that developed models have a better performance on the test sets.

Similar articles

Cited by

References

    1. Birkhead GS, Klompas M, Shah NR. Public health surveillance using electronic health records: rising potential to advance public health. Front Public Health Serv Sys Res. 2015;4(5):25–32. doi: 10.13023/FPHSSR.0405.05. - DOI
    1. Conway PH, Mostashari F, Clancy C. The Future of Quality Measurement for Improvement and Accountability. JAMA. 2013;309(21):2215–2216. doi: 10.1001/jama.2013.4929. - DOI - PubMed
    1. Cebul RD, Love TE, Jain AK, Hebert CJ. Electronic health records and quality of diabetes care. N Engl J Med. 2011;365:825–833. doi: 10.1056/NEJMsa1102519. - DOI - PubMed
    1. Yoon D, Park MY, Choi NK, et al. Detection of adverse drug reaction signals using an electronic health records satabase: Comparison of the Laboratory Extreme Abnormality Ratio (CLEAR) algorithm. Clin Pharmacol Ther. 2012;91:467–74. doi: 10.1038/clpt.2011.248. - DOI - PubMed
    1. Hebert C, Du H, Peterson LR, Robicsek A. Electronic health record-based detection of risk factors for Clostridium difficile infection relapse. Infect Control Hospital Epidemiol. 2013;34:407–414. doi: 10.1086/669864. - DOI - PubMed

LinkOut - more resources