Automatic abstraction of imaging observations with their characteristics from mammography reports

Similar documents
Automatic Identification & Classification of Surgical Margin Status from Pathology Reports Following Prostate Cancer Surgery

Boundary identification of events in clinical named entity recognition

Keeping Abreast of Breast Imagers: Radiology Pathology Correlation for the Rest of Us

A Simple Pipeline Application for Identifying and Negating SNOMED CT in Free Text

Text mining for lung cancer cases over large patient admission data. David Martinez, Lawrence Cavedon, Zaf Alam, Christopher Bain, Karin Verspoor

Background. Structured Thyroid Ultrasound Reports 1/8/2018. Clear Communication Improves Management

S. Murgo, MD. Chr St-Joseph, Mons Erasme Hospital, Brussels

Pathologic outcomes of coarse heterogeneous calcifications detected on mammography

Modeling Annotator Rationales with Application to Pneumonia Classification

A Study of Abbreviations in Clinical Notes Hua Xu MS, MA 1, Peter D. Stetson, MD, MA 1, 2, Carol Friedman Ph.D. 1

Semantic Alignment between ICD-11 and SNOMED-CT. By Marcie Wright RHIA, CHDA, CCS

CHAPTER 2 MAMMOGRAMS AND COMPUTER AIDED DETECTION

Mammogram Analysis: Tumor Classification

Text Mining of Patient Demographics and Diagnoses from Psychiatric Assessments

Mammogram Analysis: Tumor Classification

BI-RADS Update. Martha B. Mainiero, MD, FACR, FSBI Brown University Rhode Island Hospital

Investigating the performance of a CAD x scheme for mammography in specific BIRADS categories

Query Refinement: Negation Detection and Proximity Learning Georgetown at TREC 2014 Clinical Decision Support Track

COMPARISON OF BREAST CANCER STAGING IN NATURAL LANGUAGE TEXT AND SNOMED ANNOTATED TEXT

Standardize and Optimize. Trials and Drug Development

Digital Imaging and Communications in Medicine (DICOM) Supplement 50: Mammography Computer-Aided Detection SR SOP Class

Imaging in breast cancer. Mammography and Ultrasound Donya Farrokh.MD Radiologist Mashhad University of Medical Since

Breast Imaging Lexicon

CLAMP-Cancer an NLP tool to facilitate cancer research using EHRs Hua Xu, PhD

Mammography is a most effective imaging modality in early breast cancer detection. The radiographs are searched for signs of abnormality by expert

Amammography report is a key component of the breast

Factuality Levels of Diagnoses in Swedish Clinical Text

Leveraging Expert Knowledge to Improve Machine-Learned Decision Support Systems

NAÏVE BAYES CLASSIFIER AND FUZZY LOGIC SYSTEM FOR COMPUTER AIDED DETECTION AND CLASSIFICATION OF MAMMAMOGRAPHIC ABNORMALITIES

Pneumonia identification using statistical feature selection

8/31/2016 HIDING IN PLAIN SITE, ARCHITECTURAL DISTORTIONS AND BREAST ASYMMETRIES ARCHITECTURAL DISTORTIONS ARCHITECTURAL DISTORTIONS

On-time clinical phenotype prediction based on narrative reports

Clinical Narrative Analytics Challenges

Chapter 12 Conclusions and Outlook

BI-RADS Categorization As a Predictor of Malignancy 1

Incorporation of Imaging-Based Functional Assessment Procedures into the DICOM Standard Draft version 0.1 7/27/2011

Detection of architectural distortion using multilayer back propagation neural network

Leonard M. Glassman MD

Innovative Risk and Quality Solutions for Value-Based Care. Company Overview

Retrieving disorders and findings: Results using SNOMED CT and NegEx adapted for Swedish

arxiv: v1 [stat.ml] 23 Jan 2017

TeamHCMUS: Analysis of Clinical Text

A long journey to short abbreviations: developing an open-source framework for clinical abbreviation recognition and disambiguation (CARD)

ORIGINAL ARTICLE EVALUATION OF BREAST LESIONS USING X-RAY MAMMOGRAM WITH HISTOPATHOLOGICAL CORRELATION

COMPARATIVE STUDY ON FEATURE EXTRACTION METHOD FOR BREAST CANCER CLASSIFICATION

Tool Support for Cancer Lesion Tracking and Quantitative Assessment of Disease Response

BMC Medical Informatics and Decision Making 2006, 6:30

General Symptom Extraction from VA Electronic Medical Notes

ACRIN 6666 IM Additional Evaluation: Additional Views/Targeted US

TIES Cancer Research Network Y2 Face to Face Meeting U24 CA October 29 th, 2014 University of Pennsylvania

Symbolic rule-based classification of lung cancer stages from free-text pathology reports

The Radiology Aspects

Identifying Adverse Drug Events from Patient Social Media: A Case Study for Diabetes

Mammographic imaging of nonpalpable breast lesions. Malai Muttarak, MD Department of Radiology Chiang Mai University Chiang Mai, Thailand

Ad Hoc Classification of Radiology Reports

Annotating Clinical Events in Text Snippets for Phenotype Detection

Neural Network Based Technique to Locate and Classify Microcalcifications in Digital Mammograms

Erasmus MC at CLEF ehealth 2016: Concept Recognition and Coding in French Texts

Automatic Identification of Pneumonia Related Concepts on Chest x-ray Reports

Clinical Event Detection with Hybrid Neural Architecture

The Application of Image Processing Techniques for Detection and Classification of Cancerous Tissue in Digital Mammograms

Australian Journal of Basic and Applied Sciences

Name of Policy: Computer-aided Detection (CAD) Mammography

now a part of Electronic Mammography Exchange: Improving Patient Callback Rates

Diagnostic Dilemmas of Breast Imaging

Look differently. Invenia ABUS. Automated Breast Ultrasound

Since its introduction in 2000, digital mammography has become

MEDICAL POLICY SUBJECT: MAMMOGRAPHY: COMPUTER- AIDED DETECTION (CAD) POLICY NUMBER: CATEGORY: Technology Assessment

Artificial Intelligence in Breast Imaging

Learning the Fine-Grained Information Status of Discourse Entities

Building an Ensemble System for Diagnosing Masses in Mammograms

Extracting geographic locations from the literature for virus phylogeography using supervised and distant supervision methods

Clinical Trial and Evaluation of a Prototype Case-Based System for Planning Medical Imaging Work-up Strategies

Semi-Automatic Construction of Thyroid Cancer Intervention Corpus from Biomedical Abstracts

Headings: Information Extraction. Natural Language Processing. Blogs. Discussion Forums. Named Entity Recognition

Classification of benign and malignant masses in breast mammograms

Asthma Surveillance Using Social Media Data

Drug side effect extraction from clinical narratives of psychiatry and psychology patients

Memory-Augmented Active Deep Learning for Identifying Relations Between Distant Medical Concepts in Electroencephalography Reports

Building Blocks for Effective Report Communication

BI-RADS 3 category, a pain in the neck for the radiologist which technique detects more cases?

Wikipedia-Based Automatic Diagnosis Prediction in Clinical Decision Support Systems

A REVIEW ON CLASSIFICATION OF BREAST CANCER DETECTION USING COMBINATION OF THE FEATURE EXTRACTION MODELS. Aeronautical Engineering. Hyderabad. India.

Data -Driven Ontologies for an Information Extraction System from Polish Mammography Reports

An Intelligent Writing Assistant Module for Narrative Clinical Records based on Named Entity Recognition and Similarity Computation

Classification of Mammograms using Gray-level Co-occurrence Matrix and Support Vector Machine Classifier

Automatically extracting, ranking and visually summarizing the treatments for a disease

Case-based reasoning using electronic health records efficiently identifies eligible patients for clinical trials

Stepwise Knowledge Acquisition in a Fuzzy Knowledge Representation Framework

DESCRIPTION: Percentage of final reports for screening mammograms that are classified as probably benign

DOWNLOAD EASYCIE AND DATASET

EDUCATION/TRAINING (Begin with baccalaureate or other initial professional education, such as nursing, and include postdoctoral training.

Breast asymmetries in mammography: Management

Financial Disclosures

Is Probably Benign Really Just Benign? Peter R Eby, MD, FSBI Virginia Mason Medical Center Seattle, WA

Lung Cancer Diagnosis from CT Images Using Fuzzy Inference System

Retrospective study comparison of Arcadia Lab s Galileo and Parascript AccuDetect Galileo CAD systems in screening environment.

Retrospective Study Comparison of Arcadia Lab s Galileo and Parascript AccuDetect Galileo CAD systems in screening environment.

DESCRIPTION: Percentage of final reports for screening mammograms that are classified as probably benign

Session 4: Test instruments to assess interpretive performance challenges and opportunities Overview of Test Set Design and Use

Transcription:

Journal of the American Medical Informatics Association Advance Access published November 13, 2014 Research and applications Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt, 1 Jafi A Lipson, 2 Utku Senol, 3 Daniel L Rubin 4,5 1 Department of Biostatistics and Medical Informatics, Akdeniz University Faculty of Medicine, Antalya, Turkey 2 Department of Radiology, Stanford University, Stanford, California, USA 3 Department of Radiology, Akdeniz University Faculty of Medicine, Antalya, Turkey 4 Department of Radiology, Stanford University, Stanford, California, USA 5 Department of Medicine (Biomedical Informatics Research), Stanford University, Stanford, California, USA Correspondence to Dr Daniel L Rubin, Department of Radiology, Stanford University, Richard M. Lucas Center, 1201 Welch Road, Office P285, Stanford, CA 94305-5488, USA; dlrubin@stanford.edu Received 26 May 2014 Revised 11 August 2014 Accepted 4 September 2014 To cite: Bozkurt S, Lipson JA, Senol U, et al. J Am Med Inform Assoc Published Online First: [please include Day Month Year] doi:10.1136/amiajnl- 2014-003009 ABSTRACT Background Radiology reports are usually narrative, unstructured text, a format which hinders the ability to input report contents into decision support systems. In addition, reports often describe multiple lesions, and it is challenging to automatically extract information on each lesion and its relationships to characteristics, anatomic locations, and other information that describes it. The goal of our work is to develop natural language processing (NLP) methods to recognize each lesion in free-text mammography reports and to extract its corresponding relationships, producing a complete information frame for each lesion. Materials and methods We built an NLP information extraction pipeline in the General Architecture for Text Engineering (GATE) NLP toolkit. Sequential processing modules are executed, producing an output information frame required for a mammography decision support system. Each lesion described in the report is identified by linking it with its anatomic location in the breast. In order to evaluate our system, we selected 300 mammography reports from a hospital report database. Results The gold standard contained 797 lesions, and our system detected 815 lesions (780 true positives, 35 false positives, and 17 false negatives). The precision of detecting all the imaging observations with their modifiers was 94.9, recall was 90.9, and the F measure was 92.8. Conclusions Our NLP system extracts each imaging observation and its characteristics from mammography reports. Although our application focuses on the domain of mammography, we believe our approach can generalize to other domains and may narrow the gap between unstructured clinical report text and structured information extraction needed for data mining and decision support. INTRODUCTION The interpretation of mammography images is challenged by variability among radiologists in their assessment of the likelihood of malignancy given the abnormalities seen in mammography reports, and methods to improve radiologist performance are needed. 1 3 One approach to reducing radiologist variation is to standardize the vocabulary used in mammography reports. The American College of Radiology (ACR) developed the Breast Imaging-Reporting and Data System (BI-RADS). Terms called BI-RADS descriptors can be used by radiologists to describe breast density, lesion features, impression, and recommendations in mammography reports. 4 6 While adoption of BI-RADS can help to standardize the vocabulary of reports and reduce the variation in mammography reporting, it is not a decision support system (DSS) for mammography, although others have used it as a key ingredient of DSS. 7 8 For example, several works have demonstrated that statistical machine learning models such as neural networks and Bayesian networks can be built using BI-RADS descriptors and clinical data as inputs. 7 9 Although preliminary work to develop DSS for mammography using standardized vocabulary to describe the imaging features is promising, few DSS for mammography have been adopted in clinical practice, likely due to the challenge of interfacing DSS with the clinical workflow. DSS for mammography can disrupt the workflow, 7 10 12 since these systems require the radiologist to enter their observations in a separate interface, which duplicates the activity of generating the radiology report. Even if DSS are integrated into a structured reporting system, radiologists generally find it most efficient to produce reports as narrative texts rather than composing reports using structured reporting interfaces. On the other hand, recording the imaging features in unstructured narrative text prevents the use of reports directly as input to DSS, since the report content is not structured and machine-understandable. Natural language processing (NLP) techniques for information extraction could acquire the structured information from reports needed to provide the inputs to DSS, 13 assuming all pertinent imaging features are recorded in the report. In addition to providing extracted data as inputs for DSS, such NLP methods could improve interoperability among different medical information systems by providing consistent, structured data, enabling epidemiological research and clinical studies. The goal of this work is to develop and evaluate NLP methods to extract information on lesions and their corresponding imaging features (imaging observations with their modifiers) from free-text mammography reports, with the ultimate goal of providing inputs to DSS to guide radiology practice and to reduce variability in mammography interpretations. BACKGROUND Information extraction from free text is a key task for our work. A number of NLP systems have been developed for information extraction from clinical records. 14 21 In general, these systems annotate syntactic structures (e.g., sections, sentences, phrases (chunks), tokens (words), and their parts-of-speech), perform named entity recognition, map spans of text to concepts (such as drugs, pathologies, treatments, etc.) from a controlled vocabulary or ontology, and identify the negation context of named entities. 22 29 Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1 9. doi:10.1136/amiajnl-2014-003009 1

Few previous works have addressed information extraction from mammography reports. 30 34 Mammography reports are unique in several respects compared to other types of clinical narrative texts, since they contain specialized types of named entities (BI-RADS descriptors) and unique information extraction goals. Specifically, it is necessary to identify all the imaging observations and their relationships (e.g., modifiers such as imaging observation characteristics, and contexts such as negation and breast location). It is particularly important to associate the descriptions of imaging observations with the lesions to which they refer, since there are often multiple lesions in a mammogram. In this regard, mammography reports are generally not directly amenable for use in currently available medical information extraction systems, which were not developed specifically for these unique information extraction tasks. Nassif et al 32 developed a system to detect BI-RADS concepts in mammography reports, but did not consider relationships between concepts. Likewise, Jain et al used MedLEE to encode the clinical information in chest and mammogram reports to identify suspected tuberculosis and breast cancer. 31 35 Since MedLEE was not optimized for the task of correlating body locations and visual abnormalities, Sevenster et al 36 developed a system based on MedLEE, which extracts clinical findings and body locations from radiology reports and correlates them. They found that MedLEE occasionally failed to extract breast locations with multiple modifiers, such as upper inner quadrant of the left breast, although they added the BI-RADS body location phrases to MedLEE s vocabulary. They concluded that MedLEE s grammatical rules need to be updated to accommodate this type of body location; however, as MedLEE is not open source, one cannot directly modify MedLEE grammar. 36 A particularly important limitation of the NLP systems developed to date is that they do not generally recognize and disambiguate multiple lesions that are described in the same breast, and associate all of the characteristics of each lesion described in the report with the particular lesion to which they refer. 37 This is a very important problem in mammography reports, since a concept (e.g., mass ) is often used multiple times, referring to either the same or different lesions. Hence, it is important to identify whether mentions of concepts actually refer to the same lesion or to different lesions. For instance, figure 1 shows an example report describing two lesions (two different masses) having different characteristics such as margin, stability, and shape. To provide accurate input for a DSS from this kind of report, it is important for the NLP system to recognize the different masses and each of their characteristics so as to produce a complete information frame suitable for input into a DSS, as shown in the output section of figure 1. Our work tackles the challenge of recognizing mentions of breast lesions in mammography reports, extracting their relationships, and associating the extracted information with the respective lesions. Our methods recognize named entities and relations pertinent to mammography, and also recognize and link descriptions of lesions to the particular lesions to which they refer. Our approach produces a succinct, complete information frame for each lesion, its observations, characteristics, and location in a mammogram, as shown in table 1 and figure 1. Such structured summaries could be directly input into a DSS to guide physician interpretation of the imaging findings. MATERIALS AND METHODS A common strategy in developing NLP systems is to decompose the processing strategy into several subtasks, with each subtask being executed sequentially before subsequent higher-level tasks are executed. 37 We decomposed our NLP problem into a set of such subtasks, and we created a sequential processing pipeline comprising a set of sequentially executed processing modules to automatically recognize the pertinent named entities in mammography reports and the relationships among them needed to generate inputs for DSS. Our pipeline is shown in figure 2. Since a mammography report can describe multiple lesions, associating each lesion with its respective characteristics can be very challenging. We developed a strategy to tackle this challenge by using the anatomic locations of lesions to disambiguate lesions and their associated characteristics, since it is very unlikely that two abnormalities would occur in precisely the same anatomic location. Knowledge representation Our information extraction task focuses on recognizing three semantic types of named entities: imaging observations, their characteristics (modifiers of imaging observations), and anatomic locations of imaging observations. Imaging observations are terms that refer to abnormalities in the breast, such as masses and Figure 1 Example mammography report describing two different masses and the ideal output from a natural language processing system to extract the information suitable for input to a decision support system. 2 Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1 9. doi:10.1136/amiajnl-2014-003009

Table 1 Entity types and modifiers (relationships) to be extracted from mammography reports Entity type Description Example Imaging Observation Conditions, problems, etc. The breast, overall, demonstrates no focal dominant mass, architectural distortion, or suspicious calcifications. Location Anatomical structure or location of an Imaging Observation The breast, overall, demonstrates no focal dominant mass, architectural distortion, or suspicious calcifications. Location Modifiers Laterality Clock face location Depth Quadrant location Region Distance Imaging Observation Modifiers Stability Margin signal Shape signal Density signal Size signal Relates an anatomic entity to its sidedness: right, left, bilateral Relates an Imaging Observation to other information about the Imaging Observation right breast 12 o clock position anterior depth upper breast central region 1 cm from the nipple stable focal asymmetric density spiculated mass irregularly shaped mass breast tissue is largely fatty 10 6 10 mm mass calcifications. Imaging observation characteristics (also called modifiers) are terms that modify imaging observations to describe their features, such as spiculation to describe the margins of a mass in the breast. Anatomic locations are regions in which abnormalities are located, such as the upper outer quadrant of the right breast. We thus pre-defined the following semantic types of named entities and their relations: Imaging Observation, Imaging Observation Modifiers, Location, and Location Modifiers. The Location and Location Modifiers are used to associate observations with an anatomic location in the breast, represented as extracted location relationships (table 1). We used the BI-RADS controlled terminology 4 and a subset RadLex, 38 which contains terms for anatomic locations (these are not contained in BI-RADS), to provide the preferred names of all named entities. Since BI-RADS was not available in a structured format, we created a simple ontology structure of this terminology for our system ( BI-RADS ontology ). The BI-RADS ontology comprises an is-a hierarchy of entities with the following attributes: the preferred name of each term, synonyms, and acronyms (figure 3). We developed the list of synonyms and abbreviations by inspecting a set of 100 reports in addition to a predefined list compiled by experts. Processing pipeline We built our processing pipeline in the GATE (General Architecture for Text Engineering) NLP toolkit. 39 GATE is an open source framework and also provides several configurable processing resources to parse words, numbers, and punctuation (Tokenizer), and to annotate documents based on keyword lists (Gazetteer lists). In addition, GATE provides a regular expression pattern matching engine called the Java Annotation Patterns Engine ( JAPE), which is a multi-pass regular expression parser tightly integrated with GATE s document annotation representation. 39 In our pipeline (figure 2), sequential processing Figure 2 Natural language processing information extraction processing pipeline. Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1 9. doi:10.1136/amiajnl-2014-003009 3

Section segmentation This processing module segments input radiology report text into sections. Although radiology reports are free text, they are actually loosely structured in terms of report sections. Each section has its own heading in most cases (e.g., History, Findings, Impression, etc.). Section segmentation was implemented as a JAPE grammar consisting of a set of phases to annotate section headers and the subtext that belongs to those section headers. Sometimes a report may not have a section header to indicate, for example, the Findings section. To tackle that challenge, this processing component implements an additional JAPE grammar that implements logic necessary to identify the section. For example, the Findings section of mammography reports commonly starts with a sentence which describes the breast density. If there is no section header, the JAPE grammar recognizes a sentence that contains the first mention of breast density in the report as the starting point of the Findings section. In addition, sentences, noun phrases, and verb groups in each section are associated with the related section heading using JAPE grammar. Finding and linking Imaging Observation entities, modifiers, and anatomic locations for each lesion This processing module scans the noun phrases in the Findings section (since imaging observations are typically recorded in the Findings sections of radiology reports), detects BI-RADS terms to which they correspond, and produces Imaging Observation, Modifier, and Location annotations to produce information extraction frames (in examples 1 and 2 below, text representing a named entity is shown in square brackets and its semantic type is shown as a subscript). Figure 3 Breast Imaging-Reporting and Data System (BI-RADS) ontology. The is-a hierarchy is shown and the entity names are the term preferred names (synonyms are not shown). modules are executed, producing an output of structured information required for a mammography DSS. The various processing modules that accomplish each of the various subtasks comprising our system are described below, and correspond to the modules shown in figure 1. Pre-processing (Tokenizer, POS tagger, numbers/measurements/date) This module pre-processes input reports using standard GATE modules for tokenization of words, numbers and punctuation, part of speech tagging, stemming, annotation of sentences, noun phrases, verb groups, and measurements. BI-RADS onto-gazetteer This processing module processes the input text to annotate the text using terms in the BI-RADS ontology. We used the Onto-Gazetteer GATE component to access the BI-RADS ontology and to incorporate it into our pipeline. This processing module can associate the entities from a specific gazetteer list with a class in the BI-RADS ontology and is aware of mappings between lists and class IDs. We created Gazetteer lists for each class in the BI-RADS ontology that has synonyms and acronyms so that this processing module can match input text to BI-RADS terms in the ontology by both exact match and synonym match. Example 1: Information extraction of an imaging observation and its location. [The breast tissue] Location is [heterogeneously dense] ImagingObservation Example 2: Information extraction of an imaging observation, its modifier, location, and location modifier. There is [a 1 cm oval nodular density] ImagingObservation with [an obscured margin] ImagingObservation_modifier in [the right breast] Location in [the anterior depth] Location _ modifier This processing module implements several rules (using regular expressions in JAPE) to link modifiers with their entities (table 2). Determining the context of entities: negation, experiencer, and temporal status Clinical conditions can be modified by several contextual properties that are relevant for our information extraction task; ConText 40 identifies three contextual values in addition to NegEx s negation: hypothetical, historical, and experiencer. For instance, a query for patients with a diagnosis of pneumonia may return false positive records as pneumonia is mentioned but is negated (e.g., ruled out pneumonia ), is experienced by a family member (e.g., family history of pneumonia ), or occurred in the past (e.g., past history of pneumonia ). 40 We implemented ConText as a processing resource within our GATE NLP system. We used ConText to determine whether an Imaging Observation entity is negated and to determine its temporal status. For example, for input text: No significant masses or calcifications or other findings are seen in the right breast, the output of this module returns abnormalities as negated, and if it is a a previously described nodular density, the output returns the abnormality as historical. 4 Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1 9. doi:10.1136/amiajnl-2014-003009

Table 2 Sample rules for linking modifiers with Imaging Observation and Location entities Rules Example sentences Features of the entities Rule 1: Any modifier inside the text of an Imaging Observation or Location entity is added as a feature of the entity and the modifier is removed. Rule 2: If there are only stop words (e.g., in, at, of, with, etc.) or verb group(s) between an entity and its modifier, the modifier is added as a feature to the entity and the modifier is removed. Rule 3: After rule 1 and 2 are applied, if a Findings sentence contains only one entity and its modifier(s), the modifier(s) are added as the feature(s) to the entity (example 6) and the modifier is removed. There is [a 1 cm size oval shape nodular density] ImagingObservation There is [a 1 cm size oval shape nodular density] ImagingObservation with an [obscured margin margin ] ImagingObservation_modifier in [the right laterality breast] Location in [the anterior depth depth ] Location_modifier [The mass] ImagingObservation in the central region of the left breast is [stable stability ] ImagingObservation_modifier ImagingObservation: class=mass, size=1 cm (value=1, unit=cm), shape=oval ImagingObservation: class=mass, size=1 cm (value=1, unit=cm), shape=oval, margin=obscured Location: class=breast, laterality=right, depth=anterior ImagingObservation: class=mass, stability=stable Imaging Observation Location relationship extraction To associate the Location information with the Imaging Observations, we wrote JAPE grammar rules (table 4) for Patterns 1 and 2 below and added Location information as a feature for every Imaging Observation. The default value of Location was assigned as not identified. Pattern 1: First, any Location modifier inside the Imaging Observation annotation (it is annotated as a noun phrase) is added as a Location feature of the annotation (example 3). Example 3: [a [right] Location breast enhancing focus tissue] ImagingObservation: class = Mass, location = breast (Laterality=right) Pattern 2: If the Location modifier is not within the Imaging Observation entity text, to associate an Imaging Observation with Location information and to add it as a feature to an Imaging Observation entity, we allow a maximum of four stop words (SW) and one verb group (VG) between Location and Imaging Observation concepts for expressions below (table 3). We developed a simple co-reference resolution module for only Imaging Observation entities. First we tag demonstrative pronouns: this, that, these, and those as Pronoun and the Imaging Observation which contains any of these pronouns as co-referent while keeping the Imaging Observation features (class, modifier, etc.). Then, for a three sentence span, we find the Imaging Observations and their co-referents by using JAPE grammar to define the expressions below: Expression 1: {Sentence contains {ImagingObservation.class=A} & {Co-referent.class=A}} Expression 2: {Sentence contains {ImagingObservation.class=A}} {Sentence contains {Co-referent.class=A}}? (optional) {Sentence contains {Co-referent.class=A}} Where these patterns matched in the Findings section of a report, the features of the co-referent were cloned to its Imaging Observation pair, and co-referents were removed. This process helps to avoid duplications in information extraction from the report. 42 Co-reference resolution The same Imaging Observations can be referenced in different sentences, or information related to one Imaging Observation can be divided among different sentences. It is important to identify whether several Imaging Observation mentions actually refer to the same Imaging Observation. For example, in the sentences in example 4, this mass and 5 mm oval mass refer to the same Imaging Observation. In other words, these are co-referential, and resolution of such co-references serves the critical role of linking related information. 41 Example 4: In the central right breast, 5 cm posterior to the nipple, there is [a 5 mm oval mass] Imaging Observation. [This mass] Co-referent is well circumscribed. Extraction of abnormalities for each report To extract Imaging Observations that describe multiple abnormalities, it is essential to relate Imaging Observations to the particular abnormality. Our solution approach to this problem is to use the Location of lesions to identify them uniquely. Gooch and Roudsari 42 considered the set of all mentions, created subsets according to mention class, and within each subset, compared pairs of mentions in document order. Similarly, taking the set of all Imaging Observations, we create subsets of Imaging Observation entities according to their class (determined from the BI-RADS ontology, e.g., mass, calcification, associated findings, other findings, special cases), and within each subset, compare pairs of Imaging Observation entities in document order. In addition to the class of Imaging Observations, our Table 3 Grammars to detect Location and Imaging Observation relations: stop word (SW), verb group (VG), zero or one (?) Grammar Example sentences Features of the entities {Location} {SW}?{SW}?{VG}?{SW}? {SW}? {Imaging Observation} {Imaging Observation} {SW}?{SW}?{VG}?{SW}? {SW}? {Location} right breast demonstrates fibroglandular tissue. In the central right breast approximately 3 o clock there is a 5 mm oval mass. There is a right breast enhancing focus in the 8 o clock position posteriorly. ImagingObservation: class=breast_density, LocatedIn=breast: laterality=right ImagingObservation: class=mass, LocatedIn=breast: laterality=right, clock_face_location=3 o clock, region=central ImagingObservation: class=mass, LocatedIn=breast: laterality=right, clock_face_location=8 o clock, depth=central Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1 9. doi:10.1136/amiajnl-2014-003009 5

Table 4 Accuracy in extracting complete descriptions of breast lesions (Imaging Observation, characteristic, and location) Results (counts using our system) Lesions (without characteristics) Lesions (with characteristics) Size Stability Shape Margin Density Full match 780 723 184 452 232 165 104 Partial match 57 Non-match 17 FN 35 FP 1FP 5FN 6FP 18 FN 2FN 4FN 1FN Precision rate (%) 95.7 88.7 99.4 98.6 Recall rate (%) 97.8 90.7 97.3 96.1 99.1 97.6 99 Total (gold standard set) 797 797 190 476 234 169 105 Results reported in terms of full, partial, and non-matches to the actual information frames determined in the gold standard of 797 lesions. FN, false negative; FP, false positive. comparison criterion for the Imaging Observation is Location. For example, the first ImagingObservations.class=Mass is tested against all succeeding occurrence of ImagingObservations. class=mass entities in the text, the second against the third, fourth, etc. Comparisons are made based on Location information for each Imaging Observation. At the end of this process we identified only the Imaging Observations which differ from each other (example 5 in table 5). In some reports (example 6 in table 6), the Location information of two observations is exactly the same, but the sizes of the masses are different. Therefore as a second criterion, if Location information is the same and both observations have size features, their sizes are compared (i.e., we use size to disambiguate the lesions in the same location). Dataset In order to develop and evaluate our system, we selected a set of 500 mammography reports from a hospital report database, which reports were created using a structured reporting application (PenRad, Buffalo, Minnesota, USA). We selected the reports in this collection so that it had a balanced representation of the different BI-RADS classification codes (varying from benign to highly suspicious for malignancy). The reporting system produced reports with both structured data fields and free text; thus, it is a suitable gold standard, since the free-text report has associated radiologist-vetted structured data fields. However, the correspondence between structured data fields in the database and the free-text report content may not always agree, because the structured database is populated by the initial draft report created by the radiologist, while corrections or edits to the report during report review had been done only in the text report. The corresponding structured database was thus not updated if edits to the text report were made. We dealt with this limitation by having an expert mammographer review the reports and the corresponding database entries in those reports Table 5 Example 5 Text [A right breast enhancing focus] in [the 8 o clock position posteriorly] which measures [5 4 mm (AP ML)]. In [the 12 o clock position of the right breast at mid depth] there is [an oval 9 3 mm (AP ML) mass]. Imaging Observations 1. Imaging Observation: class=mass, size=5 4 mm 2. Location: laterality=right, clock face location=8, depth=posterior 3. Imaging Observation: class=mass, size=9 3 mm 4. Location: laterality=right, clock face location=12, depth=middle that we used for evaluating our system, and recording the correct structured data entries for each report. We used 200 of the 500 reports for iterative development and refinement of the NLP system; the 300 not used for developing the system were selected and held out as an independent test set for the final evaluation of the NLP system. A mammographer reviewed these 300 reports and database entries to establish the correct set of database entries for use in comparing against the study results. Thus, this 300-report dataset comprised the gold standard for our evaluation. Evaluation We used the 300 mammography report held-out dataset to evaluate our NLP system. We compared the information extraction frames produced by our system with the structured report information determined by the radiologist who reviewed the reports. We evaluated the completeness of our information extraction by calculating the precision, recall, and the F measure for extracting the correct information frames (each lesion and its characteristics). In doing these assessments, we counted complete matches (all data in the information frame is correct), partial matches (at least one modifier of an imaging observation describing a lesion is incorrect or absent), and non-matches (description of a lesion is completely missed). We calculated precision as the number of correctly identified items divided by the total number of items identified by our system, and recall as the number of correctly identified items divided by the total number of correct items). 43 We calculated the F measure as F=(2 precision recall)/ Table 6 Example 6 Text There is [a 2.5 cm round mass with a circumscribed margin] in [the right breast at 12 o clock in the anterior depth]. There also is [a 1.5 cm oval mass with a circumscribed margin] in [the right breast at 12 o clock in the anterior depth]. There is [a 1.8 cm round mass with a circumscribed margin] in the [left breast in the anterior depth central to the nipple]. Compared to previous films this mass is increased in size. There also is [a 1.4 cm oval mass with an obscured margin] in the [left breast in the anterior depth of the inferior region]. Compared to previous films this mass is increased in size. Imaging observations 1. Imaging Observation: class=mass, size=2.5 cm, location= (laterality=right, clock face location=12, depth= anterior) 2. Imaging Observation: class=mass, size=1.5 cm, location= (laterality=right, clock face location=12, depth=anterior) 3. Imaging Observation: class=mass, size=1.5 cm, location= (laterality=left, depth=anterior, region=central), stability=increase 4. Imaging Observation: class=mass, size=1.4 cm, location= (laterality=left, depth=anterior, region=inferior), stability=increase 6 Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1 9. doi:10.1136/amiajnl-2014-003009

(precision+recall). 43 We computed these three metrics at the level of exact match. In addition, true positive and false positive rates were recorded. RESULTS The gold standard set of 300 reports contained a total of 797 reported lesions. In 102 of the reports, the expert mammographer noted a deficiency in the structured data entries in the reporting application database (cases having duplicated, missed, or deficient reporting of lesions and their characteristics due to edits that were made to the text reports but were not reflected in the reporting application s database). Only the structured data entries determined by the expert mammographer were used in assessing the performance of our NLP system. Our system detected 815 lesions; 780 were true positives, 35 were false positives, and 17 were false negatives (lesions reported in the gold standard set that were not detected by the system). There were 57 partially matched cases among the detected lesions. In addition, there were 12 cases where calcification type was not detected correctly and 8 cases where breast density was not detected. For the goal of perfect match information frame extraction, the precision of extracting the mentions of Imaging Observations with their modifiers by our system was 94.9, recall was 90.9, and the F measure was 92.8. Table 4 shows the final results for full, partial, and non-match cases with precision and recall rates based on number of lesions in the 300-mammography report gold standard. DISCUSSION In this paper, we describe an NLP system to extract structured imaging observation information from mammography reports to produce an output comprising each imaging observation with its associated characteristics and anatomic locations. This information is necessary to drive a DSS, such as a Bayesian network model, to predict the probability of malignancy of a given suspicious finding in a mammogram. 7 8 The recall and precision of our system for achieving perfect information extraction for this task were 92.9% and 89.4%, respectively, which are reasonably good results, and if generalizable to other reports, could make our approach useful for information extraction for driving decision support in mammography with narrative reporting. We are currently conducting an evaluation study to assess the impact of imperfect recall and precision of information extraction on the output of a DSS. A few studies have addressed extracting information from mammography reports, including string matching methods and rule-based methods. 30 34 36 Among them, the study most similar to ours was conducted by Nassif et al in 2009. 32 The results with Nassif s system are inferior to ours, likely since it does not identify whether multiple mentions of an imaging observation refer to the same lesion or to a different lesion. Therefore, it overestimates the number of lesions present and cannot link the characteristics of a lesion to the particular lesion being described; the system extracts imaging observations based on how many times the imaging observations were mentioned in the text. Our system recognizes the context of the mentions of lesion characteristics by identifying the different lesions in the same (or different) breast and connecting the characteristics to them. In addition, Nassif s system only detects negation for each imaging observation but does not check the temporality of an imaging observation. That system only extracts size information for the last imaging observation in the report. Although relationships were not extracted, their system gave reasonably accurate results in extracting the characteristics of the imaging observations. 32 It is difficult to establish a large, robust gold standard corpus of reports for our work because most radiology reports are unstructured, and it is costly and time-consuming to read reports and record the structured data by hand. We thus leveraged a large sample of data coded by a structured reporting application, in which the data entry by the radiologist is structured as part of the reporting workflow, and the final output of the report is free text. This enabled us to create a large dataset for building and refining our system, although there were some limitations to this approach. First, since the bulk of the text is automatically generated by the reporting system, the variability in narrative reporting style was not represented, which could make our results seem better than when applied to reports not generated using this system. On the other hand, the radiologists edited this system-produced text in finalizing the report in approximately a third of cases, which likely added a substantial amount of variety to the language of the reports. Another limitation is that if the radiologist edits the text report, corresponding changes are not made in the structured database (a functional limitation of the particular reporting system). Upon review of the structured reporting application s database of structured entries and its text reports, we found a number of cases in which the database lacked some findings. Therefore, to establish the dataset for use in our final evaluation, we asked an expert mammographer to review all 300 reports in our testing report corpus and confirm the accuracy of the extracted findings recorded in the report database. Although a larger test set might be desirable, we believe this was sufficiently large for this initial evaluation of our system. Future studies with larger and more diverse reports (specifically not generated by a structured reporting application) would be helpful to confirm the accuracy and generalizability of our approach. Another limitation in our work is that although the regular expressions we used in our system seemed to work well for detecting BI-RADS descriptors, this approach is specific to BI-RADS and might not work as well if radiologists use other types of terminologies in reporting mammograms. BI-RADS is a standardized terminology and recommended for reporting all mammography, so this is not likely a substantial limitation. In addition, our system is extensible, and deficiencies in entity recognition could likely be overcome by adding the appropriate regular expressions to our processing modules. Our system was developed in the modular GATE platform, and we encapsulated the domain-specific processing in JAPE processing resources, whose regular expressions are compact and human-readable. We thus believe that such extensions are practical in the future. Although we included the BI-RADS ontology in our system, in our system we used it simply for term lookup, and we did not leverage the subsumption relations in the ontological terminology structure. We did use the ontological structure as a visual resource in creating our system; displaying the class hierarchy of the ontology was helpful in defining how to assign classes to terms in the text as we developed our JAPE grammars. We are considering extensions to our system in order to use the ontological knowledge structure in the future to enable the system to extract much more meaningful information about the text. In radiology reports, some sentences do not describe findings, and eliminating them can improve system performance. Our system does section segmentation to eliminate non-relevant sections of the report, except for the Findings section which contains the key sentences that are generally relevant to our Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1 9. doi:10.1136/amiajnl-2014-003009 7

information extraction task. Of course, even within Findings section of reports, there may be irrelevant sentences. However, the information extraction logic we developed is designed to identify and extract specific information about imaging findings. Thus, sentences about past examinations, relations between findings, or other non-relevant information in text will not be extracted, since it will not meet the criteria for our information extraction rule logic. There were some cases in which our system failed to extract the correct information from mammography reports. To detect negation and temporality, we used the ConText plug-in, which in a recent study was reported to have a 97% recall and precision for negation, and 67.4 82.5% recall and 74.2 94.3% precision for temporality (when assigning historical or hypothetical values). 44 Although ConText was generally effective in our system, we still experienced some problems detecting negation and temporality of concepts. For example, in the sentence, A coarse calcification noted in this region was removed, our system extracted the calcification as being present, because was removed was not accepted as a negation term. In another example sentence, The previously noted dense spiculated mass at the 8 o clock position appears considerably less apparent being decreased in both density and size, our system failed to extract the mass because it is mentioned in a historical context. In addition, our system does not detect historical mentions of masses or description of their stability over time, which accounts for failing to extract this example correctly. We could improve the system in the future by adding a module to recognize historical lesions and description of their stability. Another limitation of our system is related to the span of sentences considered when searching for imaging observations and their modifiers. For co-reference resolution, we assume that a span of three consecutive sentences is sufficient to link all co-referents and their modifiers to their related imaging observations. However, in a few reports we saw that the co-referents could be found in more than a three-sentence span. In addition, our system does not extract the count or plurality information for abnormalities. Therefore for the text, several nodular densities or two 8 mm intramammary lymph node were reported, our system extracted only one density and one lymph node. An additional limitation is that some of the assumptions we make about report structure in building our system may not generalize to reports created at other institutions; for example, it is possible that section headers may be omitted. To overcome this in our system, if there is no section header, a sentence that contains the first mention of breast density is accepted and considered to be the starting point of the Findings section in the report. We make this assumption because there is a convention in mammography reporting that the first sentence in the Findings section describes the breast density; however, it is possible that in some practices this may not be the case, and in such cases, our system might fail to recognize Findings section if the section header is omitted. Our system is extensible, and such variances could be accommodated in future versions if we encounter them in practice. Our approach is unique in a few respects. A common approach in information extraction systems is to create rules that find related concepts based on a 5- or 10-word span. 32 On the other hand, we developed rules based on noun and verb phrases to link the characteristics to the concepts. Our system also recognizes and extracts the contexts and relationships of imaging observations to enable us to extract information about each lesion in mammography reports, and to associate the characteristics of each lesion to the respective lesions described in the reports. This is critical for DSS, since such systems provide decision support on a per lesion basis. In addition, the rules we developed to build our system focus on concepts and relationships (imaging observations, locations, size, etc.) that are used in radiology domains other than breast imaging. Thus, we believe that our system may be extended to other domains, and we plan to expand our work to other types of radiology reports. Our information extraction system is based on rule-based methods. It is possible we could have created similar functionality using statistical methods. Some clinical NLP work is shifting from rule-based or expert-based approaches to statistical machine-learning methods, and investigating various useful features and appropriate algorithms. 28 37 Both approaches have their advantages and disadvantages. 28 33 45 46 Rule-based methods are easier to interpret, whereas statistical methods may be more robust to noise in the unstructured data (assuming they are built from large, representative corpora). In general, rulebased systems are more useful in narrow domains where human involvement is both essential and available. 47 Since mammography reporting is a narrow domain, we developed our system using rule-based methodology. Combining both rule-based and statistical methods in the future could provide even better results and generalizability. Finally, since our ultimate goal is to extract information needed for decision support, we are planning to use the output of our system as an input for a mammography DSS and to evaluate its effectiveness. We will also study the impact of imperfect precision and recall of our system on the quality of decision support delivered by a DSS. CONCLUSION There is a tremendous amount of data in unstructured free-text mammography reports, and extracting the structured content about the imaging observations and characteristics of each lesion reported could be useful to enable decision support. We developed an NLP system to extract information in mammography reports needed for input into DSS. The main contribution of this paper is the text processing methods that extract imaging observations and their attributes from mammography reports in the setting of multiple breast lesions, relating the observations and characteristics to each lesion. We believe our approach will help to narrow the gap between unstructured clinical report text and structured data capture needed for decision support and other applications. In future work we will incorporate our NLP system into a DSS. Collaborators Hakan Bulu, Department of Radiology, Stanford University, Stanford, California, USA. Contributors DLR conceived and directed the project. SB and DLR designed and evaluated the NLP system, analyzed and evaluated the data, wrote the paper, and had full access to all of the data in the study and take responsibility for the integrity of the data and the accuracy of the data analysis. JAL and US helped to create the gold standard dataset and the annotation schema. JAL and US contributed equally. All authors reviewed and approved the manuscript. Funding This work was supported by a Scientific and Technological Research Council of Turkey grant (number 2214-A). Competing interests None. Provenance and peer review Not commissioned; externally peer reviewed. REFERENCES 1 Elmore JG, Wells CK, Lee CH, et al. Variability in radiologists interpretations of mammograms. N Engl J Med 1994;331:1493 9. 8 Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1 9. doi:10.1136/amiajnl-2014-003009

2 Jiang Y, Nishikawa RM, Schmidt RA, et al. Potential of computer-aided diagnosis to reduce variability in radiologists interpretations of mammograms depicting microcalcifications. Radiology 2001;220:787 94. 3 Kerlikowske K, Grady D, Barclay J, et al. Variability and accuracy in mammographic interpretation using the American College of Radiology Breast Imaging Reporting and Data System. J Natl Cancer Inst 1998;90:1801 9. 4 Liberman L, Menell JH. Breast imaging reporting and data system (BI-RADS). Radiol Clin North Am 2002;40:409 30, v. 5 Park CS, Lee JH, Yim HW, et al. Observer agreement using the ACR Breast Imaging Reporting and Data System (BI-RADS)-ultrasound, First Edition (2003). Korean J Radiol 2007;8:397 402. 6 Burnside ES, Sickles EA, Bassett LW, et al. The ACR BI-RADS experience: learning from history. J Am Coll Radiol 2009;6:851 60. 7 Burnside ES, Rubin DL, Fine JP, et al. Bayesian network to predict breast cancer risk of mammographic microcalcifications and reduce number of benign biopsy results: initial experience. Radiology 2006;240:666 73. 8 Burnside ES, Davis J, Chhatwal J, et al. Probabilistic computer model developed from clinical data in national mammography database format to classify mammographic findings. Radiology 2009;251:663 72. 9 Ayer T, Chen Q, Burnside ES. Artificial neural networks in mammography interpretation and diagnostic decision making. Comput Math Methods Med 2013;2013:832509. 10 Wu Y, Giger ML, Doi K, et al. Artificial neural networks in mammography: application to decision making in the diagnosis of breast cancer. Radiology 1993;187:81 7. 11 Taylor P. Decision support for image interpretation: a mammography workstation. Inf Process Med Imaging 1995;3:227 38. 12 Stivaros SM, Gledson A, Nenadic G, et al. Decision support systems for clinical radiological practice towards the next generation. Br J Radiol 2010;83:904 14. 13 Demner-Fushman D, Chapman WW, McDonald CJ. What can natural language processing do for clinical decision support? J Biomed Inform 2009;42:760 72. 14 Friedman C, Alderson PO, Austin JH, et al. A general natural-language text processor for clinical radiology. J Am Med Inform Assoc 1994;1:161 74. 15 Chapman WW, Haug PJ. Comparing expert systems for identifying chest x-ray reports that support pneumonia. Proc AMIA Symp 1999:216 20. 16 Savova GK, Masanz JJ, Ogren PV, et al. Mayo clinical Text Analysis and Knowledge Extraction System (ctakes): architecture, component evaluation and applications. J Am Med Inform Assoc 2010;17:507 13. 17 Zeng QT, Goryachev S, Weiss S, et al. Extracting principal diagnosis, co-morbidity and smoking status for asthma research: evaluation of a natural language processing system. BMC Med Inform Decis Mak 2006;6:30. 18 Crowley RS, Castine M, Mitchell K, et al. caties: a grid based system for coding and retrieval of surgical pathology reports and tissue specimens in support of translational research. J Am Med Inform Assoc 2010;17:253 64. 19 Denny JC, Irani PR, Wehbe FH 3rd, et al. The KnowledgeMap project: development of a concept-based medical school curriculum database. AMIA Annu Symp Proc 2003:195 9. 20 Hahn U, Romacker M, Schulz S. MEDSYNDIKATE a natural language system for the extraction of medical information from findings reports. Int J Med Inform 2002;67:63 74. 21 Aronson AR, Lang FM. An overview of MetaMap: historical perspective and recent advances. J Am Med Inform Assoc 2010;17:229 36. 22 Lacson R, Sugarbaker N, Prevedello LM, et al. Retrieval of radiology reports citing critical findings with disease-specific customization. Open Med Inform J 2012;6:28 35. 23 Rubin D, Wang D, Chambers DA, et al. Natural language processing for lines and devices in portable chest x-rays. AMIA Annu Symp Proc 2010;2010:692 6. 24 Do BH, Wu AS, Maley J, et al. Automatic retrieval of bone fracture knowledge using natural language processing. J Digit Imaging 2013;26:709 13. 25 Cheng LT, Zheng J, Savova GK, et al. Discerning tumor status from unstructured MRI reports completeness of information in existing reports and utility of automated natural language processing. J Digit Imaging 2010;23:119 32. 26 Yetisgen-Yildiz M, Gunn ML, Xia F, et al. A text processing pipeline to extract recommendations from radiology reports. J Biomed Inform 2013;46: 354 62. 27 Yetisgen-Yildiz M, Gunn ML, Xia F, et al. Automatic identification of critical follow-up recommendation sentences in radiology reports. AMIA Annu Symp Proc 2011;2011:1593 602. 28 Meystre SM, Savova GK, Kipper-Schuler KC, et al. Extracting information from textual documents in the electronic health record: a review of recent research. Yearb Med Inform 2008;35:128 44. 29 Bashyam V, Taira RK. Indexing anatomical phrases in neuro-radiology reports to the UMLS 2005AA. AMIA Annu Symp Proc 2005:26 30. 30 Burnside E, Rubin D, Strasberg H. Automated indexing of mammography reports using linear least squares fit. International congress; 14th Computer assisted radiology and surgery; Cars 2000: Comput Assist Radiol Surg 2000;1214:449 54. 31 Jain NL, Friedman C. Identification of findings suspicious for breast cancer based on natural language processing of mammogram reports. Proc AMIA Annu Fall Symp 1997:829 33. 32 Nassif H, Woods R, Burnside E, et al. Information Extraction for Clinical Data Mining: A Mammography Case Study. 2009 IEEE International Conference on Data Mining Workshops (ICDMW 2009); 2009:37 42. 33 Mykowiecka A, Marciniak M, Kupsc A. Rule-based information extraction from patients clinical data. J Biomed Inform 2009;42:923 36. 34 Esuli A, Marcheggiani D, Sebastiani F. An enhanced CRFs-based system for information extraction from radiology reports. J Biomed Inform 2013;46: 425 35. 35 Jain NL, Knirsch CA, Friedman C, et al. Identification of suspected tuberculosis patients based on natural language processing of chest radiograph reports. Proc AMIA Annu Fall Symp 1996:542 6. 36 Sevenster M, van Ommering R, Qian Y. Automatically correlating clinical findings and body locations in radiology reports using MedLEE. J Digit Imaging 2012;25:240 9. 37 Nadkarni PM, Ohno-Machado L, Chapman WW. Natural language processing: an introduction. J Am Med Inform Assoc 2011;18:544 51. 38 Langlotz CP. RadLex: a new method for indexing online educational materials. Radiographics 2006;26:1595 7. 39 Cunningham H, Tablan V, Roberts A, et al. Getting more out of biomedical documents with GATE s full lifecycle open source text analytics. PLoS Comput Biol 2013;9:e1002854. 40 Harkema H, Dowling JN, Thornblade T, et al. ConText: an algorithm for determining negation, experiencer, and temporal status from clinical reports. J Biomed Inform 2009;42:839 51. 41 Zheng J, Chapman WW, Crowley RS, et al. Coreference resolution: a review of general methodologies and applications in the clinical domain. J Biomed Inform 2011;44:1113 22. 42 Gooch P, Roudsari A. Lexical patterns, features and knowledge resources for coreference resolution in clinical notes. J Biomed Inform 2012;45:901 12. 43 Maynard D, Peters W, Li Y. Metrics for evaluation of ontology-based information extraction. International World Wide Web Conference; Edinburgh, UK 2006. 44 Meystre S, Haug PJ. Natural language processing to extract medical problems from electronic clinical documents: performance evaluation. J Biomed Inform 2006;39:589 99. 45 Liu F, Weng C, Yu H. Natural language processing, electronic health records, and clinical research. In: Richesson RL, Andrews JE. eds. Clinical research informatics. London: Springer, 2012:293 310. 46 Friedman C, Rindflesch TC, Corn M. Natural language processing: state of the art and prospects for significant progress, a workshop sponsored by the National Library of Medicine. J Biomed Inform 2013;46:765 73. 47 Sarawagi S. Information extraction. Foundations and trends in databases 2008;1:261 377. Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1 9. doi:10.1136/amiajnl-2014-003009 9