Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2010 Jul-Aug;17(4):375-82.
doi: 10.1136/jamia.2009.001412.

Evaluation of a generalizable approach to clinical information retrieval using the automated retrieval console (ARC)

Affiliations

Evaluation of a generalizable approach to clinical information retrieval using the automated retrieval console (ARC)

Leonard W D'Avolio et al. J Am Med Inform Assoc. 2010 Jul-Aug.

Abstract

Reducing custom software development effort is an important goal in information retrieval (IR). This study evaluated a generalizable approach involving with no custom software or rules development. The study used documents "consistent with cancer" to evaluate system performance in the domains of colorectal (CRC), prostate (PC), and lung (LC) cancer. Using an end-user-supplied reference set, the automated retrieval console (ARC) iteratively calculated performance of combinations of natural language processing-derived features and supervised classification algorithms. Training and testing involved 10-fold cross-validation for three sets of 500 documents each. Performance metrics included recall, precision, and F-measure. Annotation time for five physicians was also measured. Top performing algorithms had recall, precision, and F-measure values as follows: for CRC, 0.90, 0.92, and 0.89, respectively; for PC, 0.97, 0.95, and 0.94; and for LC, 0.76, 0.80, and 0.75. In all but one case, conditional random fields outperformed maximum entropy-based classifiers. Algorithms had good performance without custom code or rules development, but performance varied by specific application.

PubMed Disclaimer

Conflict of interest statement

Competing interests: None.

Figures

Figure 1
Figure 1
Current processes of clinical information retrieval (IR) versus those proposed in the design of the automated retrieval console (ARC).
Figure 2
Figure 2
Automated retrieval console (ARC) menu, showing the various ARC interfaces.
Figure 3
Figure 3
A screen shot of the Judge interface. The annotation instructions shown in the ‘Help Information’ window is populated as part of the creation of a new project.
Figure 4
Figure 4
A graphical representation of the study design. CRC, colorectal cancer; CRF, conditional random field; MaxEnt, maximum entropy; NLP, natural language processing.

Similar articles

Cited by

References

    1. U.S. Department of Health and Human Services About healthy people, 2009. http://www.healthypeople.gov/About/
    1. Committee on Comparative Effectiveness Research Prioritization Initial priorities for comparative effectiveness research. Washington DC: Institute of Medicine, 2009
    1. Bates D, Kuperman G, Wang S, et al. Ten commandments for effective clinical decision support: Making the practice of evidence-based medicine a reality. J Am Med Inform Assoc 2003;10:523–30 - PMC - PubMed
    1. Murphy S, Churchill S, Bry L, et al. Instrumenting the health care enterprise for discovery research in the genomic era. Genome Res 2009;19:1675–81 - PMC - PubMed
    1. Berg M, Goorman E. The contextual nature of medical information. Int J Med Inform 1999;56:51–60 - PubMed

Publication types