Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2024 Feb 16;10(5):e26434.
doi: 10.1016/j.heliyon.2024.e26434. eCollection 2024 Mar 15.

Use of noisy labels as weak learners to identify incompletely ascertainable outcomes: A Feasibility study with opioid-induced respiratory depression

Affiliations

Use of noisy labels as weak learners to identify incompletely ascertainable outcomes: A Feasibility study with opioid-induced respiratory depression

Alvin D Jeffery et al. Heliyon. .

Abstract

Objective: Assigning outcome labels to large observational data sets in a timely and accurate manner, particularly when outcomes are rare or not directly ascertainable, remains a significant challenge within biomedical informatics. We examined whether noisy labels generated from subject matter experts' heuristics using heterogenous data types within a data programming paradigm could provide outcomes labels to a large, observational data set. We chose the clinical condition of opioid-induced respiratory depression for our use case because it is rare, has no administrative codes to easily identify the condition, and typically requires at least some unstructured text to ascertain its presence.

Materials and methods: Using de-identified electronic health records of 52,861 post-operative encounters, we applied a data programming paradigm (implemented in the Snorkel software) for the development of a machine learning classifier for opioid-induced respiratory depression. Our approach included subject matter experts creating 14 labeling functions that served as noisy labels for developing a probabilistic Generative model. We used probabilistic labels from the Generative model as outcome labels for training a Discriminative model on the source data. We evaluated performance of the Discriminative model with a hold-out test set of 599 independently-reviewed patient records.

Results: The final Discriminative classification model achieved an accuracy of 0.977, an F1 score of 0.417, a sensitivity of 1.0, and an AUC of 0.988 in the hold-out test set with a prevalence of 0.83% (5/599).

Discussion: All of the confirmed Cases were identified by the classifier. For rare outcomes, this finding is encouraging because it reduces the number of manual reviews needed by excluding visits/patients with low probabilities.

Conclusion: Application of a data programming paradigm with expert-informed labeling functions might have utility for phenotyping clinical phenomena that are not easily ascertainable from highly-structured data.

Keywords: Classification; Electronic health records; Machine learning; Medical informatics; Phenotype.

PubMed Disclaimer

Conflict of interest statement

The authors declare the following financial interests/personal relationships which may be considered as potential competing interests: Alvin D. Jeffery reports financial support was provided by 10.13039/100006537Vanderbilt University School of Nursing. Alvin D. Jeffery reports a relationship with 10.13039/100000133Agency for Healthcare Research and Quality that includes: funding grants. Alvin D. Jeffery reports a relationship with 10.13039/100006093Patient-Centered Outcomes Research Institute that includes: funding grants. Alvin D. Jeffery reports a relationship with 10.13039/100000738US Department of Veterans Affairs that includes: employment and funding grants. Alvin D. Jeffery reports a relationship with 10.13039/100006108National Center for Advancing Translational Sciences that includes: funding grants. Michael E. Matheny and Ruth M. Reeves also have employment and funding grants from the US Department of Veterans Affairs. If there are other authors, they declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Figures

Fig. 1
Fig. 1
Graphical representation of research methods.
Fig. 2
Fig. 2
Flow diagram illustrating the number of patients and visits present at each phase of cohort processing. *Note: The number of unique patients in the Manually-Reviewed Test Set (702) is smaller than the sum of the 2 preceding boxes (717) because those boxes were sampled at the visit-level instead of patient-level.
Fig. 3
Fig. 3
Comparison of OIRD predicted probabilities from the Generative model with manually-adjudicated labels in Validation Set.
Fig. 4
Fig. 4
Comparison of OIRD predicted probabilities from the Discriminative model with manually-adjudicated labels in Validation Set.
Fig. 5
Fig. 5
Comparison of predicted probabilities between Generative and Discriminative models with final case/control status denoted – all visits.
Fig. 6
Fig. 6
Comparison of predicted probabilities between Generative and Discriminative models with final case/control status denoted – with visits determined to be a Control with full agreement on manual review are removed.

Update of

Similar articles

References

    1. Bastarache L., Brown J.S., Cimino J.J., et al. Developing real-world evidence from real-world data: Transforming raw data into analytical datasets. Learn Health Syst. 2022;6(1) 10.1002/lrh2.10293published Online First: 20211014. - PMC - PubMed
    1. Li Q., Melton K., Lingren T., et al. Phenotyping for patient safety: algorithm development for electronic health record based automated adverse event and medical error detection in neonatal intensive care. J. Am. Med. Inf. Assoc. : JAMIA. 2014;21(5):776–784. 10.1136/amiajnl-2013-001914published Online First: 20140108. - PMC - PubMed
    1. Newton K.M., Peissig P.L., Kho A.N., et al. Validation of electronic medical record-based phenotyping algorithms: results and lessons learned from the emerge network. J. Am. Med. Inf. Assoc. : JAMIA. 2013;20(e1):e147–e154. 10.1136/amiajnl-2012-000896published Online First: 2013/03/28. - PMC - PubMed
    1. Overby C.L., Pathak J., Gottesman O., et al. A collaborative approach to developing an electronic health record phenotyping algorithm for drug-induced liver injury. J. Am. Med. Inf. Assoc. : JAMIA. 2013;20(e2):e243–e252. 10.1136/amiajnl-2013-001930published Online First: 2013/07/11. - PMC - PubMed
    1. Alzoubi H., Alzubi R., Ramzan N., et al. A review of automatic phenotyping approaches using electronic health records. Electronics-Switz. 2019;8(11) ARTN 1235, 10.3390/electronics8111235published Online First.

LinkOut - more resources