Automatically extracting, ranking and visually summarizing the treatments for a disease

Similar documents
Analyzing the Semantics of Patient Data to Rank Records of Literature Retrieval

Summarizing Drug Information in Medline Citations

Stepwise Knowledge Acquisition in a Fuzzy Knowledge Representation Framework

Erasmus MC at CLEF ehealth 2016: Concept Recognition and Coding in French Texts

Query Refinement: Negation Detection and Proximity Learning Georgetown at TREC 2014 Clinical Decision Support Track

Biomedical resources for text mining

Kalpana Raja, PhD 1, Andrew J Sauer, MD 2,3, Ravi P Garg, MSc 1, Melanie R Klerer 1, Siddhartha R Jonnalagadda, PhD 1*

Semi-Automatic Construction of Thyroid Cancer Intervention Corpus from Biomedical Abstracts

A Study of Abbreviations in Clinical Notes Hua Xu MS, MA 1, Peter D. Stetson, MD, MA 1, 2, Carol Friedman Ph.D. 1

A Simple Pipeline Application for Identifying and Negating SNOMED CT in Free Text

How preferred are preferred terms?

HHS Public Access Author manuscript Stud Health Technol Inform. Author manuscript; available in PMC 2015 July 08.

A FRAMEWORK FOR CLINICAL DECISION SUPPORT IN INTERNAL MEDICINE A PRELIMINARY VIEW Kopecky D 1, Adlassnig K-P 1

Mining Medline for New Possible Relations of Concepts

BMC Medical Informatics and Decision Making 2011, 11:3

EMBASE Find quick, relevant answers to your biomedical questions

Case-based reasoning using electronic health records efficiently identifies eligible patients for clinical trials

Seeking Informativeness in Literature Based Discovery

A Comparison of Collaborative Filtering Methods for Medication Reconciliation

The Role of Knowledge in Conceptual Retrieval: A Study in the Domain of Clinical Medicine

Data Mining in Bioinformatics Day 4: Text Mining

38 Int'l Conf. Bioinformatics and Computational Biology BIOCOMP'16

Management of new onset atrial fibrillation McNamara R L, Bass E B, Miller M R, Segal J B, Goodman S N, Kim N L, Robinson K A, Powe N R

PROCEDURAL APPROACH TO MITIGATING CONCURRENTLY APPLIED CLINICAL PRACTICE GUIDELINES

TeamHCMUS: Analysis of Clinical Text

Chapter 12 Conclusions and Outlook

PubMed Tutorial Author: Gökhan Alpaslan DMD,Ph.D. e-vident

Situated Question Answering in the Clinical Domain: Selecting the Best Drug Treatment for Diseases

Curriculum Vitae. Degree and date to be conferred: Masters in Computer Science, 2013.

An Intelligent Writing Assistant Module for Narrative Clinical Records based on Named Entity Recognition and Similarity Computation

Selecting a research method

Comparing ICD9-Encoded Diagnoses and NLP-Processed Discharge Summaries for Clinical Trials Pre-Screening: A Case Study

Building Causal Graphs from Medical Literature and Electronic Medical Records

Assessing the role of a medication-indication resource in the treatment relation extraction from clinical text

Schema-Driven Relationship Extraction from Unstructured Text

Extracting geographic locations from the literature for virus phylogeography using supervised and distant supervision methods

The Impact of Directionality in Predications on Text Mining

Automatic Identification & Classification of Surgical Margin Status from Pathology Reports Following Prostate Cancer Surgery

READ-BIOMED-SS: ADVERSE DRUG REACTION CLASSIFICATION OF MICROBLOGS USING EMOTIONAL AND CONCEPTUAL ENRICHMENT

Cocktail Preference Prediction

IBM Research Report. Automated Problem List Generation from Electronic Medical Records in IBM Watson

Atigeo at TREC 2012 Medical Records Track: ICD-9 Code Description Injection to Enhance Electronic Medical Record Search Accuracy

Perceived similarity and visual descriptions in content-based image retrieval

Building a Knowledge Base to Support a Digital Library

Annotation and Retrieval System Using Confabulation Model for ImageCLEF2011 Photo Annotation

SAGE. Nick Beard Vice President, IDX Systems Corp.

Rate of Heart failure guideline adherence in a tertiary care center in India after accounting for the therapeutic contraindications.

Lessons Extracting Diseases from Discharge Summaries

Searching Literature Using PubMed. Dr Armen Torchyan, MD, MPH Dr Rufaidah Dabbagh, MBBS, MPH October, 2011

Elsevier ClinicalKey TM FAQs

Text mining for lung cancer cases over large patient admission data. David Martinez, Lawrence Cavedon, Zaf Alam, Christopher Bain, Karin Verspoor

Toward a Unified Representation of Findings in Clinical Radiology

General Symptom Extraction from VA Electronic Medical Notes

RESEARCH PROJECT. Comparison of searching tools and outcomes of different providers of the Medline database (OVID and PubMed).

Knowledge networks of biological and medical data An exhaustive and flexible solution to model life sciences domains

Generative Content Models for Structural Analysis of Medical Abstracts

Wikipedia-Based Automatic Diagnosis Prediction in Clinical Decision Support Systems

Semantic Alignment between ICD-11 and SNOMED-CT. By Marcie Wright RHIA, CHDA, CCS

PhenDisco: a new phenotype discovery system for the database of genotypes and phenotypes

How does the heart work? The heart is muscle whose main function is a pump; to push blood the rest of your body.

Not all NLP is Created Equal:

Chapter 2. Knowledge Representation: Reasoning, Issues, and Acquisition. Teaching Notes

How can Natural Language Processing help MedDRA coding? April Andrew Winter Ph.D., Senior Life Science Specialist, Linguamatics

A Study on a Comparison of Diagnostic Domain between SNOMED CT and Korea Standard Terminology of Medicine. Mijung Kim 1* Abstract

Memory-Augmented Active Deep Learning for Identifying Relations Between Distant Medical Concepts in Electroencephalography Reports

Supplementary Figure 1

HBML: A Representation Language for Quantitative Behavioral Models in the Human Terrain

Using Scripts to help in Biomedical Text Interpretation

How to do a quick search for evidence

ECG Beat Recognition using Principal Components Analysis and Artificial Neural Network

Evaluation of Clinical Text Segmentation to Facilitate Cohort Retrieval

Clinical Narratives Context Categorization: The Clinician Approach using RapidMiner

Automatic Indexing and Retrieving Context-Specific Evidence to Support Physician Decision Making at the Point of Care

Identifying Diagnostic Studies in MEDLINE: Reducing the Number Needed to Read

AN ONTOLOGICAL APPROACH TO REPRESENTING AND REASONING WITH TEMPORAL CONSTRAINTS IN CLINICAL TRIAL PROTOCOLS

Identifying anatomical concepts associated with ICD10 diseases

KOS Design for Healthcare Decision-making Based on Consumer Criteria and User Stories

Interval Likelihood Ratios: Another Advantage for the Evidence-Based Diagnostician

How to Create Better Performing Bayesian Networks: A Heuristic Approach for Variable Selection

Heiner Oberkampf. DISSERTATION for the degree of Doctor of Natural Sciences (Dr. rer. nat.)

Mapping the UMLS Semantic Network into General Ontologies

Weighted Ontology and Weighted Tree Similarity Algorithm for Diagnosing Diabetes Mellitus

Ambiguity of Human Gene Symbols in LocusLink and MEDLINE: Creating an Inventory and a Disambiguation Test Collection

Asthma Surveillance Using Social Media Data

Text Mining of Patient Demographics and Diagnoses from Psychiatric Assessments

Quality of Life Analysis for Patients Receiving Heart Implants David Nfodjo &Christiana Petrou, University of Louisville, Louisville, KY

Animal Disease Event Recognition and Classification

Pilot Study: Clinical Trial Task Ontology Development. A prototype ontology of common participant-oriented clinical research tasks and

Formalizing UMLS Relations using Semantic Partitions in the context of task-based Clinical Guidelines Model

A Method for Analyzing Commonalities in Clinical Trial Target Populations

Lung Cancer Concept Annotation from Spanish Clinical Narratives

Predicting Breast Cancer Survivability Rates

A Graphic User Interface for the Glyco-MGrid Portal System for Collaboration among Scientists

Modeling Annotator Rationales with Application to Pneumonia Classification

SEMANTICS-BASED CALORIE CALCULATOR. A Paper Submitted to the Graduate Faculty of the North Dakota State University of Agriculture and Applied Science

Classification of Smoking Status: The Case of Turkey

Distillation of Knowledge from the Research Literatures on Alzheimer s Dementia

EVIDENCE CENTRE REPORT

PONTE: A Context-Aware Approach for Automated Clinical Trial Protocol Design

CADTH Therapeutic Review

Transcription:

Automatically extracting, ranking and visually summarizing the treatments for a disease Prakash Reddy Putta, B.Tech 1,2, John J. Dzak III, BS 1, Siddhartha R. Jonnalagadda, PhD 1 1 Division of Health and Biomedical Informatics, Northwestern University Feinberg School of Medicine, Chicago, IL 2 Department of Management and Information Systems, University of Illinois, Chicago, IL Abstract Clinicians are expected to have up-to-date and broad knowledge of disease treatment options for a patient. Online health knowledge resources contain a wealth of information. However, because of the time investment needed to disseminate and rank pertinent information, there is a need to summarize the information in a more concise format. Our aim of the study is to provide clinicians with a concise overview of popular treatments for a given disease using information automatically computed from Medline abstracts. We analyzed the treatments of two disorders Atrial Fibrillation and Congestive Heart Failure. We calculated the precision, recall, and f-scores of our two ranking methods to measure the accuracy of the results. For Atrial Fibrillation disorder, maximum f-score for the New Treatments weighing method is 0.611, which occurs at 60 treatments. For Congestive Heart Failure disorder, maximum f-score for the New Treatments weighing method is 0.503, which occurs at 80 treatments. Introduction Clinicians are expected to have up-to-date and broad knowledge of disease treatment options for a patient. Online health knowledge resources contain a wealth of information. However, because of the time investment needed to disseminate and rank pertinent information, there is a need to summarize the information in a more concise format. In a 1985 seminal study, Covell et al. 1 observed that physicians raised two questions for every three patients seen in an outpatient setting. In 70% of the cases, these questions were not answered. More recent research 2 has produced similar results, with little improvement compared to Covell et al. s findings. Based on Ely et al. s taxonomy and their studies with clinicians at the point of care 2-4, the major information types needed at the point of care are treatment and diagnosis. Treatment questions (drugs and other therapy) account for about 44% of the questions asked and diagnosis-related questions constitute about 38% of the questions asked. The goal of this research is to provide clinicians with a concise overview of different treatments for a given disease. Given this information, a clinician can efficiently research common or emerging treatments while skipping over the unimportant treatments. In this study, we explored the feasibility of extracting treatments from Medline abstracts for Congestive Heart Failure and Atrial Fibrillation and providing an interactive display of the treatments through a web-interface. Medline abstracts have been used recently to answer clinical questions by extracting individual sentences 5. They have also been used to graphically display treatments and disorders as nodes in a graph, where the salient treatments are found using network centrality measures 6. In this paper, we are studying an alternative format of summarizing therapeutic options where the clinician users can enter their disorder of interest in natural language or code, see the treatments ranked by frequency across different year ranges (although they might even be ranked using network centrality measures across different year ranges), and also compare different treatments or combinations of them. Methods The methods are divided into 2 parts: A) description of the system that generates interactive display of the treatments for a user-chosen treatment; and B) system evaluation.

A. System Architecture The system to automatically generate treatment displays is built as a pipeline that combines the following informatics tools and methods: 1) a query processing tool using UMLS Metathesaurus 7 for extracting concepts, 2) Semantic Medline Database 8 for extracting treatment relationships, 3) a term-frequency based algorithm to eliminate non-specific treatment names, 4) a normalized-frequency based ranking of the remaining treatments, and 5) a graphical interface to interactively learn about the treatments. Figure 1 depicts the system architecture and flow. Figure 1. System Architecture 1. Query Processing As shown in Figure 1, the query for a clinical topic (e.g., congestive heart failure) is initially processed with a UIMA-based (Unstructured Information Management Architecture 9 ) concept extraction pipeline that maps concepts in narrative text to the UMLS Metathesaurus. The components in the pipeline are: Tokenization or splitting the query into individual token adapted from the OpenNLP suite 10 ; Lexical normalization or converting words to a canonical form the Lexical Variant Generation (LVG) terminology from the UMLS metathesaurus is compressed by i) converting the terms to lowercase; ii) removing the terms where the normalized word has more than one token; and iii) removing the terms that have the same base form. UMLS Metathesaurus lookup performed using a well-known efficient algorithm called Aho-Corasick string matching algorithm 11 that loads the normalized tokens and their substrings as the individual states of the corresponding finite state machine, where the transitions between the different states represent the different terms formed by the original tokens. Concept screening select only the UMLS concepts in the query that are members of the semantic groups disorder (Abnormality, Dysfunction, Disease or Syndrome, Finding, Injury or Poisoning, Pathologic Function, and Sign or Symptom 12 ). 2. Treatment extraction from Medline abstracts The treatments for the disorder extracted in the previous step is obtained by querying the Semantic Medline Database. Semantic Medline Database (SemMedDB) 8 is a repository of semantic predications (subject-predicateobject triples) extracted by SemRep 13, a semantic interpreter of biomedical text. The query we use searches SemMedDB for sentences where the object of the sentence is related to disorder and the predicate of the sentence is TREATS (e.g. radiofrequency ablation treats atrial fibrillation). Algorithm 2 of our previous work 14 details how the queries are created to retrieve the relevant semantic predications. Additionally the query is constrained to be novel which eliminates predications that have a generic argument as determined by hierarchical depth in the Metathesaurus[2]. The query then yields a set of treatments for the disorder (example: congestive heart failure), mentioned in Medline abstracts.

3. Eliminating non-specific treatment names When treatments were extracted from the Semantic Medline Database, many broadly named procedures like Therapeutic Procedure and Adjustment Action were also extracted. These procedures, being unimportant to a clinician, had to be removed while displaying keeping the relevant treatments. Some of them were removed using a feature present in SemMedDB that prunes concepts according to their relative granularity in the concept hierarchy. In addition, we also calculated the number of abstracts in which the combination of procedure (e.g. Adjustment Action ) and disorder (e.g. Congestive Heart Failure ) appears in Medline compared with the number of abstracts in which the treatment itself appears in Medline. The treatments, for which the former number was relatively low compared to the later number (< 1%) was categorized as a broadly named procedures and hence these procedures were removed from the data set. 4. Normalized-frequency based ranking There are multiple ways in which the treatments might be ordered. We chose to first normalize the treatments by their presence in Medline abstracts across individual year ranges (1980-1985, 1986-1990, 1991-1995, 1996-2000, 2001-2005, 2006-2010, and 2011-2013) and construct a linear combination of the these normalized frequencies with one combination giving more weight to the recent years, the other giving equal weights and also providing the option in the interface (discussed in next section) for the clinician user to choose their own weights. Normalization across individual year ranges was done by using the formula given below. where, E min represents the minimum value across the individual year range; E max represents the maximum value across the individual year range. If a clinician is interested in treatments that are often used in recent years, we rank the treatments by giving weights of 1, 2, 3, 4, 5, 6, and 7 for the respective year-ranges. This would result in ranking of the treatments in which more importance is given to the treatments that appeared in the recent years. The first of our weighing method was called New Treatments Weighing Method. In the New Treatments Weighing Method, we gave weights of 1, 2, 3, 4, 5, 6, and 7 to the year ranges. This will result in ranking of the treatments in which the treatments that appeared recently are given more importance. The second weighing method was called Established Treatments Weighing Method. In the Established Treatments Weighing Method, we gave equal weights to all the year ranges, i.e. weights of 1 are given to all the year ranges. This weighing method results in ranking of the treatments in which all the treatments that appeared are given equal importance. 5. Graphical interface While building the interface, we tried to convey simplicity and provide a reasonable starting point for clinicians to begin their information gathering for a given disease treatment. Building the first iteration of the tool, we started by identifying use cases a clinician would have when starting to gather treatment information for a particular disease. In the first use case we identified, clinicians search for novel disease treatments in scientific journals. In the second use case, clinicians search for established disease treatments in scientific journals. In the third use case, clinicians search are more interested in treatments studied during a particular timespan more than the others. Finally, in the fourth use case, the clinician would compare the number of individual disease treatments mentions with their intersections to find popular combinations of treatments.

B. System Evaluation For the evaluation study, we used a case study approach by assessing the output of the knowledge summary system on the treatment of two conditions: atrial fibrillation (a topic relatively well understood) and congestive heart failure (a topic less covered, but is an increasingly complex vast field with knowledge from huge literature). For atrial fibrillation, we compared the rankings of the system with the practice guideline Management of Patients With Atrial Fibrillation (Compilation of 2006 ACCF/AHA/ESC and 2011 ACCF/AHA/HRS Recommendations). 15 For atrial fibrillation, we compared the rankings of the system with the practice guideline 2013 ACCF/AHA Guideline for the Management of Heart Failure: Executive Summary: A Report of the American College of Cardiology Foundation/American Heart Association Task Force on Practice Guidelines. 16 For both these disorders, we report the precision, recall and f-score values for top-10 treatments, top-20 treatments, and so on until top-100 treatments. Precision is the ratio of the number of treatments in the guideline and extracted by the system and the number of all the treatment extracted by the system. Recall is the ratio of the number of treatments in the guideline and extracted by the system and the number of all the treatments in the gold standard. F- score is the harmonic mean of precision and recall. Results Atrial Fibrillation After we extracted the treatments for Atrial Fibrillation from SemMedDB, 623 treatments were mined. Out of these, several treatments were generalized procedures which would not be of any use to a clinician. By using e- utilities tool and a threshold frequency of 1% we cleaned the data of these non-specific treatments. On performing the cleaning, 360 of these were removed and 263 treatments remained. On these 263 treatments, we used our two ranking methods explained in the methods section. New Treatments Weighing method: The top 5 treatments that were generated using this method along with their UMLS Concept Unique Identifiers are Ablation (C0547070), Electric Counter shock (C0013778), anticoagulation (C0003281), Stroke prevention (C1277289), and Cardiac ablation (C0162563). The precision-recall curve for this weighing method for the top 100 treatments and the F-score values are displayed in Figure 2 and Table 1. Precision vs Recall for New Treatments of Atrial Fibrillation 1 0.8 Precision 0.6 0.4 0.2 0 0 0.2 0.4 0.6 0.8 1 Recall Figure 2. Precision vs Recall for New Treatments of Atrial Fibrillation

Table 1. Recall, Precision and F-scores for New Treatments of Atrial Fibrillation Number of Recall Precision F-score treatments 1 0.000 1.000 0.000 10 0.099 0.700 0.173 20 0.239 0.850 0.374 30 0.324 0.767 0.455 40 0.394 0.700 0.505 50 0.493 0.700 0.579 60 0.563 0.667 0.611 70 0.606 0.614 0.610 80 0.634 0.563 0.596 90 0.676 0.533 0.596 100 0.718 0.510 0.596 Established Treatments Weighing method: The top 5 treatments that were generated using this method are Electric Counter shock (C0013778), Ablation (C0547070), anticoagulation (C0003281), Amiodarone (C0002598), and Stroke prevention (C1277289). The precision-recall curve for this weighing method for the top 100 treatments and the F-score values are displayed in Figure 3 and Table 2. Precision vs Recall for Established Treatments of Atrial Fibrillation Precision 1 0.8 0.6 0.4 0.2 0 0 0.2 0.4 0.6 0.8 1 Recall Figure 3. Precision vs Recall for Established Treatments of Atrial Fibrillation Table 2. Recall, Precision and F-scores for Established Treatments of Atrial Fibrillation Number of Recall Precision F-Square Treatments 1 0.000 1.000 0.000 10 0.127 0.900 0.222 20 0.239 0.850 0.374 30 0.338 0.800 0.475 40 0.423 0.750 0.541 50 0.479 0.680 0.562

60 0.549 0.650 0.595 70 0.592 0.600 0.596 80 0.634 0.563 0.596 90 0.690 0.544 0.609 100 0.732 0.520 0.608 Congestive Heart Failure After we extracted the treatments for Congestive Heart Failure from SemMedDB, 1,608 treatments were mined. Out of these, several treatments were generalized procedures which would not be of any use to a clinician. On performing the cleaning, 1,355 of these were removed and 253 treatments remained. On these 253 treatments, we used our two ranking methods explained in the methods section. New Treatments Weighing method: The top 5 treatments that were generated using this method are Angiotensin- Converting Enzyme Inhibitors (C0003015), Adrenergic beta-antagonists (C0001645), Cardiac resynchronization therapy (C1167956), Brain natriuretic peptide (C0054015), and Digoxin (C0012265). The precision-recall curve for this weighing method for the top 100 treatments along with the F-score values are displayed below in Figure 4 and Table 3. Precision vs Recall for New Treatments of Congestive Heart Failure Precision 1 0.8 0.6 0.4 0.2 0 0 0.2 0.4 0.6 0.8 1 Recall Figure 4. Precision vs Recall for New Treatments of Congestive Heart Failure Table 3. Recall, Precision and F-scores for New Treatments of Congestive Heart Failure Number of Recall Precisio F-Square Treatments n 1 0.000 1.000 0.000 10 0.101 0.800 0.180 20 0.177 0.700 0.283 30 0.241 0.633 0.349 40 0.304 0.600 0.403 50 0.342 0.540 0.419 60 0.392 0.517 0.446 70 0.443 0.500 0.470 80 0.494 0.488 0.491 90 0.519 0.456 0.485

100 0.570 0.450 0.503 Established Treatments Weighing method: The top 5 treatments that were generated using this method are Angiotensin-Converting Enzyme Inhibitors (C0003015), Adrenergic beta-antagonists (C0001645), Digoxin (C0012265), Cardiac resynchronisation therapy (C1167956), and Milrinone (C0128513). The precision-recall curve for this weighing method for the top 100 treatments along with the F-score values are displayed below in Figure 5 and Table 4. Precision vs Recall for Established Treatments of Congestive Heart Failure Precision 1 0.8 0.6 0.4 0.2 0 0 0.2 0.4 0.6 0.8 1 Recall Figure 5. Precision vs Recall for Established Treatments of Congestive Heart Failure Table 4. Recall, Precision and F-scores for Established Treatments of Congestive Heart Failure Number of Treatments Recall Precision F-Square 1 0.000 1.000 0.000 10 0.114 0.900 0.202 20 0.152 0.600 0.242 30 0.215 0.567 0.312 40 0.278 0.550 0.370 50 0.329 0.520 0.403 60 0.367 0.483 0.417 70 0.418 0.471 0.443 80 0.430 0.425 0.428 90 0.468 0.411 0.438 100 0.494 0.390 0.436 Interface For using the tool, a clinician would follow the steps outlined below (See Figure 6 for a screenshot): 1. Clinician searches for a disease they want to gather more information about either by entering the disease name or the UMLS Concept Unique Identifier. 2. Clinician selects a treatment trend their interested in (Novel, Consistent, Custom) 3. Clinician sorts treatments (Trend matches: most to least, Treatment matches: least to most)

4. Clinician selects treatments they would like to see graphed by the number of mentions over time. Figure 6. Web interface for showing treatment trends Discussion Atrial Fibrillation: The precision and recall curve shows that as the recall increases, precision decreases. In a webinterface, one needs to display enough number of treatments such that both the precision and recall are moderately high. For the case of New Established Treatments method, the f-score of 0.611 is highest at precision of 0.667. The recall at this precision is 0.611. This cutoff point is reached at 60 treatments. This cutoff point is also applicable to the Established Treatments method. At 60 treatments, the f-score is 0.595, which is approximately as high as the largest f-score value, 0.609 for this method. Congestive Heart Failure: For the case of New Treatments method, the best cutoff point for f-score is 0.4915, which occurs at 80 treatments. Even though the highest f-score is 0.503 (which occurs at 100 treatments), 80 treatments would be a better cutoff since there is not much difference in the F-scores between these two and selecting the 100 treatments would mean just going through more number of treatments. The recall and precision at this cutoff are 0.494 and 0.488 respectively, which are moderately high. For the case of Established Treatments method, we can observe the maximum f-score (0.443) is reached at 70 treatments. The recall and precision values at this cutoff are 0.418 and 0.471 respectively. For the two disorders we have discussed, we can observe that a cutoff of 70 treatments would result in a moderately high precision and recall. In our interface, we are able to display as many treatments as there are for the disorder; however, text mining based interfaces that display fewer treatments might be missing few important treatments. Limitations The analysis was performed on only two diseases Atrial Fibrillation and Congestive Heart Failure. Although this might be sufficient to generalize the system for cardiology, more experiments are needed to generalize it for all the disorders. The threshold value of 1% that was used as a part of data cleaning is arbitrary decided by observing various data sets. Even though this value was helpful in removing several generalized procedures, few generalized procedures such as Hospitalization and Secondary Prevention are still not filtered. In the future, we will incorporate this ratio

threshold values and will perform analysis on treatment procedures of various diseases and come up with an optimized value for this. Another limitation is that SemRep like any text mining system misses few relationships (resulting in lower recall for our system) and falsely assigns relationships (resulting in lower precision). In a previous study, the average precision of SemRep for predications on the treatment of four diseases was 73% (recall not calculated as it was difficult to determine the false negatives). 17 Moreover, SemRep operates at a sentence-level, which means that when there is an implicit relationship between treatments and a disease (such as the disease mentioned in one sentence and the treatment mentioned using an anaphora to the disease in another), the system would not be able to extract it. When extracting the treatments, the same treatments are sometimes repeated with different names. For example, anti-coagulation and Anti-Coagulation therapy refers to the same treatment, but they are being listed separately because of the challenges with normalization of these named-entities. Another example is that of Warfarin and Warfarin Therapy. A related example is Defibrillator and Cardioversion which are synonyms according to UMLS Metathesaurus. The practice guidelines only used the term Cardioversion and in our extracted treatment names Defibrillator and Cardioversion are listed separately. Because the same treatment is being displayed under different names, the data for the number of times a treatment has appeared will not be accurate and the clinician user will also have a suboptimal experience. Conclusion and Future work For Atrial Fibrillation disorder, maximum f-score for the New Treatments weighing method is 0.611, which occurs at 60 treatments. For Congestive Heart Failure disorder, maximum f-score for the New Treatments weighing method is 0.503, which occurs at 100 treatments. By observing the precision, recall and f-scores of the ranking methods we proposed, we can infer that the extraction and ranking schema we proposed might be a good solution for displaying the treatment procedures. The analysis we have done was only on two diseases, Atrial Fibrillation and Congestive Heart Failure. In future, we will be performing similar kind of analysis on more number of diseases. By analyzing these results, we target to get a generalized optimal solution for other disorders. We discussed various limitations that reduced the accuracy of the results generated. In future studies, we aim to come up with solutions to these limitations. Acknowledgements This work was made possible by funding from the National Library of Medicine K99/R00 LM011389. References 1. Covell DG, Uman GC, Manning PR. Information needs in office practice: are they being met? Annals of Internal Medicine 1985;103:596-9. 2. Ely JW, Osheroff JA, Chambliss ML, Ebell MH, Rosenbaum ME. Answering physicians' clinical questions: obstacles and potential solutions. J Am Med Inform Assoc 2005;12:217-24. 3. Ely JW, Osheroff JA, Gorman PN, et al. A taxonomy of generic clinical questions: classification study. Bmj 2000;321:429-32. 4. Ely JW, Osheroff JA, Maviglia SM, Rosenbaum ME. Patient-care questions that physicians are unable to answer. J Am Med Inform Assoc 2007;14:407-14. 5. Cao Y, Liu F, Simpson P, et al. AskHERMES: An online question answering system for complex clinical questions. Journal of Biomedical Informatics 2011. 6. Zhang H, Fiszman M, Shin D, Miller CM, Rosemblat G, Rindflesch TC. Degree centrality for semantic abstraction summarization of therapeutic studies. Journal of Biomedical Informatics 2011;44:830-8. 7. Bodenreider O. The Unified Medical Language System (UMLS): integrating biomedical terminology. Nucleic Acids Research 2004;32:D267-70.

8. Kilicoglu H, Shin D, Fiszman M, Rosemblat G, Rindflesch TC. SemMedDB: a PubMed-scale repository of biomedical semantic predications. Bioinformatics 2012;28:3158-60. 9. Apache. Unstructured Information Management Architecture. http://uimaapacheorg2012. 10. Apache. OpenNLP. http://opennlpapacheorg/2012. 11. Aho AV, Corasick MJ. Efficient string matching: an aid to bibliographic search. Commun ACM 1975;18:333-40. 12. McCray AT, Burgun A, Bodenreider O. Aggregating UMLS semantic types for reducing conceptual complexity. Stud Health Technol Inform 2001;84:216-20. 13. Rindflesch TC, Fiszman M. The interaction of domain knowledge and linguistic structure in natural language processing: interpreting hypernymic propositions in biomedical text. J Biomed Inform 2003;36:462-77. 14. Jonnalagadda SR, Del Fiol G, Medlin R, et al. Automatically extracting sentences from Medline citations to support clinicians information needs. Journal of the American Medical Informatics Association 2012. 15. Anderson JL, Halperin JL, Albert NM, et al. Management of patients with atrial fibrillation (compilation of 2006 ACCF/AHA/ESC and 2011 ACCF/AHA/HRS recommendations): a report of the American College of Cardiology/American Heart Association Task Force on Practice Guidelines. Journal of the American College of Cardiology 2013;61:1935-44. 16. Yancy CW, Jessup M, Bozkurt B, et al. 2013 ACCF/AHA guideline for the management of heart failure: executive summary: a report of the American College of Cardiology Foundation/American Heart Association Task Force on practice guidelines. Circulation 2013;128:1810-52. 17. Fiszman M, Demner-Fushman D, Kilicoglu H, Rindflesch TC. Automatic summarization of MEDLINE citations for evidence-based medical treatment: a topic-oriented evaluation. J Biomed Inform 2009;42:801-13.