A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis
- PMID: 33323251
- DOI: 10.1016/S2589-7500(19)30123-2
A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis
Erratum in
-
Correction to Lancet Digital Health 2019; 1: e271-97.Lancet Digit Health. 2019 Nov;1(7):e334. doi: 10.1016/S2589-7500(19)30160-8. Epub 2019 Oct 9. Lancet Digit Health. 2019. PMID: 33345807 No abstract available.
Abstract
Background: Deep learning offers considerable promise for medical diagnostics. We aimed to evaluate the diagnostic accuracy of deep learning algorithms versus health-care professionals in classifying diseases using medical imaging.
Methods: In this systematic review and meta-analysis, we searched Ovid-MEDLINE, Embase, Science Citation Index, and Conference Proceedings Citation Index for studies published from Jan 1, 2012, to June 6, 2019. Studies comparing the diagnostic performance of deep learning models and health-care professionals based on medical imaging, for any disease, were included. We excluded studies that used medical waveform data graphics material or investigated the accuracy of image segmentation rather than disease classification. We extracted binary diagnostic accuracy data and constructed contingency tables to derive the outcomes of interest: sensitivity and specificity. Studies undertaking an out-of-sample external validation were included in a meta-analysis, using a unified hierarchical model. This study is registered with PROSPERO, CRD42018091176.
Findings: Our search identified 31 587 studies, of which 82 (describing 147 patient cohorts) were included. 69 studies provided enough data to construct contingency tables, enabling calculation of test accuracy, with sensitivity ranging from 9·7% to 100·0% (mean 79·1%, SD 0·2) and specificity ranging from 38·9% to 100·0% (mean 88·3%, SD 0·1). An out-of-sample external validation was done in 25 studies, of which 14 made the comparison between deep learning models and health-care professionals in the same sample. Comparison of the performance between health-care professionals in these 14 studies, when restricting the analysis to the contingency table for each study reporting the highest accuracy, found a pooled sensitivity of 87·0% (95% CI 83·0-90·2) for deep learning models and 86·4% (79·9-91·0) for health-care professionals, and a pooled specificity of 92·5% (95% CI 85·1-96·4) for deep learning models and 90·5% (80·6-95·7) for health-care professionals.
Interpretation: Our review found the diagnostic performance of deep learning models to be equivalent to that of health-care professionals. However, a major finding of the review is that few studies presented externally validated results or compared the performance of deep learning models and health-care professionals using the same sample. Additionally, poor reporting is prevalent in deep learning studies, which limits reliable interpretation of the reported diagnostic accuracy. New reporting standards that address specific challenges of deep learning could improve future studies, enabling greater confidence in the results of future evaluations of this promising technology.
Funding: None.
Copyright © 2019 The Author(s). Published by Elsevier Ltd. This is an Open Access article under the CC BY 4.0 license. Published by Elsevier Ltd.. All rights reserved.
Comment in
-
Human versus machine in medicine: can scientific literature answer the question?Lancet Digit Health. 2019 Oct;1(6):e246-e247. doi: 10.1016/S2589-7500(19)30124-4. Epub 2019 Sep 24. Lancet Digit Health. 2019. PMID: 33323246 No abstract available.
Similar articles
-
Artificial intelligence performance in detecting tumor metastasis from medical radiology imaging: A systematic review and meta-analysis.EClinicalMedicine. 2020 Dec 25;31:100669. doi: 10.1016/j.eclinm.2020.100669. eCollection 2021 Jan. EClinicalMedicine. 2020. PMID: 33392486 Free PMC article.
-
Unassisted Clinicians Versus Deep Learning-Assisted Clinicians in Image-Based Cancer Diagnostics: Systematic Review With Meta-analysis.J Med Internet Res. 2023 Mar 2;25:e43832. doi: 10.2196/43832. J Med Internet Res. 2023. PMID: 36862499 Free PMC article. Review.
-
Diagnostic accuracy of artificial intelligence for detecting gastrointestinal luminal pathologies: A systematic review and meta-analysis.Front Med (Lausanne). 2022 Nov 4;9:1018937. doi: 10.3389/fmed.2022.1018937. eCollection 2022. Front Med (Lausanne). 2022. PMID: 36405592 Free PMC article.
-
Beyond the black stump: rapid reviews of health research issues affecting regional, rural and remote Australia.Med J Aust. 2020 Dec;213 Suppl 11:S3-S32.e1. doi: 10.5694/mja2.50881. Med J Aust. 2020. PMID: 33314144
-
Reviewing the relationship between machines and radiology: the application of artificial intelligence.Acta Radiol Open. 2021 Feb 9;10(2):2058460121990296. doi: 10.1177/2058460121990296. eCollection 2021 Feb. Acta Radiol Open. 2021. PMID: 33623711 Free PMC article. Review.
Cited by
-
Few-shot learning for inference in medical imaging with subspace feature representations.PLoS One. 2024 Nov 6;19(11):e0309368. doi: 10.1371/journal.pone.0309368. eCollection 2024. PLoS One. 2024. PMID: 39504337 Free PMC article.
-
Diagnostic performance of deep learning for infectious keratitis: a systematic review and meta-analysis.EClinicalMedicine. 2024 Oct 18;77:102887. doi: 10.1016/j.eclinm.2024.102887. eCollection 2024 Nov. EClinicalMedicine. 2024. PMID: 39469534 Free PMC article.
-
The Integration of Radiomics and Artificial Intelligence in Modern Medicine.Life (Basel). 2024 Oct 1;14(10):1248. doi: 10.3390/life14101248. Life (Basel). 2024. PMID: 39459547 Free PMC article. Review.
-
Studies of Artificial Intelligence/Machine Learning Registered on ClinicalTrials.gov: Cross-Sectional Study With Temporal Trends, 2010-2023.J Med Internet Res. 2024 Oct 25;26:e57750. doi: 10.2196/57750. J Med Internet Res. 2024. PMID: 39454187 Free PMC article.
-
Perception of the Adoption of Artificial Intelligence in Healthcare Practices Among Healthcare Professionals in a Tertiary Care Hospital: A Cross-Sectional Study.Cureus. 2024 Sep 22;16(9):e69910. doi: 10.7759/cureus.69910. eCollection 2024 Sep. Cureus. 2024. PMID: 39439624 Free PMC article.
Publication types
MeSH terms
Grants and funding
LinkOut - more resources
Full Text Sources
Medical
