Evaluation of SNOMED Coverage of Veterans Health Administration Terms

Similar documents
A Study on a Comparison of Diagnostic Domain between SNOMED CT and Korea Standard Terminology of Medicine. Mijung Kim 1* Abstract

Can SNOMED CT replace ICD-coding?

A Descriptive Delta for Identifying Changes in SNOMED CT

Building a Diseases Symptoms Ontology for Medical Diagnosis: An Integrative Approach

Using SNOMED CT Codes for Coding Information in Electronic Health Records for Stroke Patients

Chapter 12 Conclusions and Outlook

Interface Terminologies: Facilitating Direct Entry of Clinical Data into Electronic Health Record Systems

A Study of Abbreviations in Clinical Notes Hua Xu MS, MA 1, Peter D. Stetson, MD, MA 1, 2, Carol Friedman Ph.D. 1

EDUCATION/TRAINING (Begin with baccalaureate or other initial professional education, such as nursing, and include postdoctoral training.

Semantic Alignment between ICD-11 and SNOMED-CT. By Marcie Wright RHIA, CHDA, CCS

Matching GP terms to the ICD-10-AM index

SNOMED CT and Orphanet working together

A Simple Pipeline Application for Identifying and Negating SNOMED CT in Free Text

Redwood Mednet Connecting California to Improve Patient Care 2012 Conference. Michael Stearns, MD, CPC, CFPC HIT Consultant

Evaluating E&M Coding Accuracy of GoCode as Compared to Internal Medicine Physicians and Auditors

Not all NLP is Created Equal:

Biomedical resources for text mining

Integrating Personalized Health Information from MedlinePlus in a Patient Portal

How to code rare diseases with international terminologies?

Knowledge Representation of Traditional Chinese Acupuncture Points Using the UMLS and a Terminology Model

Asthma Surveillance Using Social Media Data

FINAL REPORT Measuring Semantic Relatedness using a Medical Taxonomy. Siddharth Patwardhan. August 2003

IBM Research Report. Automated Problem List Generation from Electronic Medical Records in IBM Watson

Authoring SNOMED CT Generic Authoring Principles. Penni Hernandez and Cathy Richardson Senior Terminologists

An introduction to case finding and outcomes

Analyzing the Semantics of Patient Data to Rank Records of Literature Retrieval

Big Data Phenomics in the VA. Outline

Abstract. Introduction and Background. by Hari Nandigam MD, MSHI, and Maxim Topaz, PhD, RN, MA

Auditing SNOMED Relationships Using a Converse Abstraction Network

HHS Public Access Author manuscript Stud Health Technol Inform. Author manuscript; available in PMC 2015 July 08.

SNOMED CT: Who Needs to Know What?

38 Int'l Conf. Bioinformatics and Computational Biology BIOCOMP'16

The UMLS-CORE project: a study of the problem list terminologies used in large healthcare institutions

An Ontology for Healthcare Quality Indicators: Challenges for Semantic Interoperability

EXPLORING THE INTERNAL CONSISTENCY OF REGISTRY DATA ON STAGE OF DISEASE AT DIAGNOSIS

SAGE. Nick Beard Vice President, IDX Systems Corp.

2018 OPTIONS FOR INDIVIDUAL MEASURES: REGISTRY ONLY. MEASURE TYPE: Process

Moving Family Health History & Genetic Test Result Data into the Electronic Health Record for Clinical Decision Support

Innovative Risk and Quality Solutions for Value-Based Care. Company Overview

CHAPTER 6 DESIGN AND ARCHITECTURE OF REAL TIME WEB-CENTRIC TELEHEALTH DIABETES DIAGNOSIS EXPERT SYSTEM

WELSH INFORMATION STANDARDS BOARD

USER GUIDE. SNOMED RT Systematized Nomenclature of Medicine Reference Terminology VERSION 1.0

Rare Diseases Nomenclature and classification

A Comparison of Collaborative Filtering Methods for Medication Reconciliation

Text mining for lung cancer cases over large patient admission data. David Martinez, Lawrence Cavedon, Zaf Alam, Christopher Bain, Karin Verspoor

Lessons Extracting Diseases from Discharge Summaries

DEVELOPMENT AND EVALUATION OF A PERIOPERATIVE MEDICATION MANAGEMENT DECISION SUPPORT TOOL

The Significance of SNODENT

PubMed Tutorial Author: Gökhan Alpaslan DMD,Ph.D. e-vident

Detecting Patient Complexity from Free Text Notes Using a Hybrid AI Approach

Automatic Identification of Pneumonia Related Concepts on Chest x-ray Reports

How can Natural Language Processing help MedDRA coding? April Andrew Winter Ph.D., Senior Life Science Specialist, Linguamatics

Automatic generation of MedDRA terms groupings using an ontology

PERSONALIZED GENETIC REPORT CLIENT-REPORTED DATA PURPOSE OF THE X-SCREEN TEST

Extracting geographic locations from the literature for virus phylogeography using supervised and distant supervision methods

2017 OPTIONS FOR INDIVIDUAL MEASURES: CLAIMS ONLY. MEASURE TYPE: Process

Demonstrating Collect once, Use Many Assimilating Public Health Secondary Data Use Requirements into an Existing Domain Analysis Model

NUMERATOR: Reports that include the pt category, the pn category and the histologic grade

With Semantic Analysis from Noun Phrases to SNOMED CT and Classification Codes H. R. Straub, M. Duelli, Semfinder AG, Kreuzlingen, Switzerland

History of SETMA s Preparation for ICD-10 By James L. Holly, MD October 1, 2015

Common Errors. ClinicalTrials.gov Basic Results Database

Electronic Support for Public Health Vaccine Adverse Event Reporting System (ESP:VAERS)

Erasmus MC at CLEF ehealth 2016: Concept Recognition and Coding in French Texts

Investigating implementing CEN with HL7 V3 and SNOMED CT Final Report

Standardize and Optimize. Trials and Drug Development

Semantic Clarification of the Representation of Procedures and Diseases in SNOMED CT

Effect of (OHDSI) Vocabulary Mapping on Phenotype Cohorts

Clinical Decision Support Systems. 朱爱玲 Medical Informatics Group 24 Jan,2003

Heiner Oberkampf. DISSERTATION for the degree of Doctor of Natural Sciences (Dr. rer. nat.)

A Lexical-Ontological Resource forconsumerheathcare

The Veterans Health Administration (VHA) has undergone. equality: Electronic Quality Assessment From Narrative Clinical Reports

The Potential of SNOMED CT to Improve Patient Care. Dec 2015

COMPARISON OF BREAST CANCER STAGING IN NATURAL LANGUAGE TEXT AND SNOMED ANNOTATED TEXT

CLINICIAN-LED E-HEALTH RECORDS. (AKA GETTING THE LITTLE DATA RIGHT) Dr Heather Leslie Ocean Informatics/openEHR Foundation

A FRAMEWORK FOR CLINICAL DECISION SUPPORT IN INTERNAL MEDICINE A PRELIMINARY VIEW Kopecky D 1, Adlassnig K-P 1

Advancing methods to develop behaviour change interventions: A Scoping Review of relevant ontologies

GENERALIZABILITY AND RELIABILITY: APPROACHES FOR THROUGH-COURSE ASSESSMENTS

Copyright Warning & Restrictions

Retrieving disorders and findings: Results using SNOMED CT and NegEx adapted for Swedish

Needles in a haystack: screening and healthcare system evidence for homelessness

DoD/VA Allergen Terminology Usage Analysis

A Framework for Conceptualizing, Representing, and Analyzing Distributed Interaction. Dan Suthers

How preferred are preferred terms?

NUMERATOR: Reports that include the pt category, the pn category and the histologic grade

MMI HIT Integration, Interoperability, & Standards Group 1 Project SONMED CT November 27, 2012

WikiWarsDE: A German Corpus of Narratives Annotated with Temporal Expressions

A Model for Evaluating Interface Terminologies

Defining and Evaluating Resilience: A Performability Perspective

Truth Versus Truthiness in Clinical Data

HPV Vaccination. Steps for Increasing. in Practice. An Action Guide to Implement Evidence-based Strategies for Clinicians*

Data Harmonization Efforts Across the Cancer Staging System. Donna M. Gress, RHIT, CTR AJCC Technical Specialist

General Symptom Extraction from VA Electronic Medical Notes

Keeping Abreast of Breast Imagers: Radiology Pathology Correlation for the Rest of Us

The diagnosis of Chronic Pancreatitis

Pilot Study: Clinical Trial Task Ontology Development. A prototype ontology of common participant-oriented clinical research tasks and

Robert A. Greenes, MD, PhD

The openehr approach. Background. Approach. Heather Leslie, July 17, 2014

OBSERVATIONAL MEDICAL OUTCOMES PARTNERSHIP

Kaiser Permanente Convergent Medical Terminology (CMT)

UDHQ Project Overview and Update 7 May 2012

Transcription:

MEDINFO 2004 M. Fieschi et al. (Eds) Amsterdam: IOS Press 2004 IMIA. All rights reserved Evaluation of SNOMED Coverage of Veterans Health Administration Terms Janet FE Penz a, Steven H Brown b, John S Carter c, Peter L Elkin d, Viet N Nguyen a, Shannon A Sims a, Michael J Lincoln a a Veterans Administration Hospital and University of Utah Department of Medical Informatics, Salt Lake City, Utah b Veterans Administration Hospital, Nashville, Tennessee c Apelon,, Inc., Ridgefield, CT d Mayo Clinic, Laboratory of Biomedical Informatics, Rochester, MN Janet FE Penz, Steven H Brown, John S Carter, Peter L Elkin, Viet N Nguyen, Shannon A Sims, Michael J Lincoln Abstract Veterans Health Administration (VHA) is now evaluating use of SNOMED-CT. This paper reports the first phase of this evaluation, which examines the coverage of SNOMED-CT for problem list entries. Clinician expressions in VA problem lists are quite diverse compared to the content of the current VA terminology Lexicon. We selected a random set of 5054 narratives that were previously unresolved against the Lexicon. These narratives were mapped to SNOMED-CT using two automated tools. Experts reviewed a subset of the tools matched, partly matched, and un-matched narratives. The automated tools produced exact or partial matches for over 90% of the 5054 unresolved narratives. SNOMED-CT has promise as a coding system for clinical problems. In subsequent studies, VA will examine the coverage of SNOMED for other clinical domains, such as drugs, allergies, and physician orders. Keywords Veterans Health Administration, Terminology, SNOMED, Controlled Vocabulary, Medical Records Systems, Computerized Introduction The Veterans Health Administration (VHA) is the largest integrated health system in the United States and has a vast repository of clinical data that is currently stored as either coded or narrative text. Veterans Health Information Systems and Technology Architecture (VISTA) [1] forms the framework for VHA s Computerized Patient Record System (CPRS). VHA is now preparing to implement a national (and locally replicated) Health Data Repository (HDR) as well as re-engineered financial and demographic database systems. When fully implemented, the HDR will be a structured, patient-centric database that supports the electronic medical record for patients. It will receive, store, and send patient-specific information to and from ancillary systems and other repositories inside or outside of the enterprise. HDR will have to support the real-time or near realtime transaction analysis now supported by VISTA databases, such clinician health reminders and drug interaction checking. It will function as the base structure for broader, population-based analysis as well. VHA has now decided to standardize its clinical terminologies across the enterprise in order to accommodate an HDR of national scope. VA clinicians across the nation now enter practically all notes, orders, problems, etc using VA s Computerized Patient Record System (CPRS). As these individual CPRS systems are re-engineered to feed the HDR, data standardization becomes a primary concern. For example, VISTA problem list entries are currently resolved against the VA Lexicon Utility. This utility, in existence for 10 years, contains VA specific terminology as well as a variety of standard coding systems such as MeSH, ICD-9, and CPT. However, clinical problem expressions have proved to be considerably more diverse than Lexicon can accommodate. When clinicians fail to find a suitable code for a problem narrative, they can optionally choose their narrative as a free text problem expression. As a result, VA has been accumulating hundreds of thousands of unresolved problem narratives each year. These unresolved narratives are often more granular or more pre-coordinated (e.g., chest pain radiating to the left arm ) than the existing Lexicon content. The U.S. Federal government has recently negotiated a license for the Systematic Nomenclature of MEDicine-Clinical Terms (SNOMED-CT). This formal reference terminology is compatible with the VA s Enterprise Reference Terminology and may offer advantages over the existing VHA terminology sources such as the Lexicon. The VA is now evaluating how to use SNOMED-CT. This paper reports upon that evaluation s first phase, examining the coverage of SNOMED-CT for VA problem list terms and phrases. SNOMED-CT contains approximately 344,000 concepts, 913,000 English language descriptions and more than one million semantic relationships [2]. It is a multi-axial, hierarchical coding scheme that has been previously shown to have broad coverage of health terminologies [3,4]. Methods Unresolved problem narratives arise at all VA sites. A random sample of these narratives (URN s) comprising 5,054 terms was selected for this study. These narratives have a broad scope and include abbreviations, diagnoses, procedures, medical risk factors, information about patient involvement in research trials, symptoms, laboratory findings, etc. Some narrative phrases in- 540

volve complex, pre-coordinated concepts, for example, radiation proctitis with lower GI bleed and severe anemia. The narratives also include abbreviations, misspellings and non-standard phrases or words such as transaminitis. Therefore, not all unresolved narratives are necessarily codable. We selected a random set of 5,054 unresolved narratives and submitted each to two automated SNOMED mapping tools. One tool was provided by Apelon, Inc. as part of its TermWorks product and this searched SNOMED for concept names matching the unresolved narratives using a multi-step algorithm. The algorithm uses word-based tokenization, word normalization and stemming to find matching concepts. From the set of results returned, we used the "best match" as determined by highest number of tokens matched as an entry point into SNOMED-CT for expert reviewers to evaluate whether or not an appropriate concept describing the health factor was found in SNOMED- CT." The narratives were also submitted to the SmartAccess Vocabulary Server (SAVS; pronouced Saves ), which has been developed by Dr. Peter Elkin s laboratory at the Mayo Clinic, Rochester, MN. The SAVS system is an indexing tool with source materials using a Concept-Based Indexing Schema. The ontology of medical concepts available in the SNOMED-CT terminology provides the underlying indexing schema. The software has been extensively tested in Mayo s Usability Laboratory. SAVS provides capability for spelling correction, synonym and coding schemes and data storage. For each term we compared the output of the SNOMED mapping for both tools. The automated results were classified as complete matches when each tool produced the same SNOMED code(s). The results were classed as partial matches when the two tools produced non-congruent SNOMED code(s). If neither tool generated a SNOMED match, then the result was classified as failed. Thus the automated matching results on each narrative fell into one of these three groups. A clinician examined partial or failed matches and corrected any spelling errors and expanded acronyms or abbreviations. Subsequently several clinicians with coding experience (JFEP, MJL, SS, VN) examined samples which totaled 200 of the 603 complete matches, 400 of the 4148 partial matches, and 303 of the 303 failed matches. Because of the relatively small numbers in this group the sample was made up of all 303. Clue-5 Browser ( http://www.clinical-info.co.uk/), a look-up engine for SNOMED CT terms, was used to confirm accuracy of the tool generated SNOMED concept codes and to assess for the existence of codes in the situations in which there was either incorrect code generation or where no code at all was generated. Results The initial comparison of the SNOMED concept ID codes generated by each tool is summarized in Table 1. Table 2 shows the results of the evaluation and expert review of these samples of URN s. When the sample of 200 complete matches were manually compared to the corresponding SNOMED concept obtained using the Clue Browser the majority did demonstrate accurate assignment of codes by the automatic tools. Determinations were made by expert reviewers regarding the existence of appropriate SNOMED concepts for any elements that were not captured by the automatic tools to assess the overall adequacy of SNOMED to cover the URN terms. As expected there was a high degree of agreement in this group having URN terms for which both tools provided identical concept codes. 83% of complete matches were determined to be accurate in the coverage of the URN and with manual coding a total of 93.5% of the terms were found to have acceptable and completely encompassing SNOMED concepts. Table 1: Comparison of the degree of agreement in SNOMED concept ID codes generated for URN s by both TermWorks and SmartAccess tools. Tool output No. (%) Complete agreement 604 (11.9) Partial agreement 4148 (82.1) Failed with no output 303 (6.0) Total 5054 (100.0) Partial tool matches formed the largest category comprising 4148 of 5055 URN s. The random sample of 400 were treated similarly and yielded 46.5% accuracy by the automatic tools matching and with an additional 34% being partially accurate. Common reasons for incomplete match or partially incorrect match were abbreviations and spelling errors in a portion of a composite term that the tools were unable match. Expert review was able to capture almost all remaining terms. Thus a total of 91.3% of this sample showed accurate coverage by SNOMED concepts. The third category consisting of URN s for which neither tool could provide a SNOMED code ( failed matches ) included 303 URN s. Initial correction of spelling and abbreviation errors were necessary during expert review and manual assessment of the existence of appropriate SNOMED concepts to cover the URN terms. Although the reviewers could not resolve 104 terms, 197 of 199 terms (99%) could be accurately manually matched to SNOMED codes. Two remaining terms were found to have inadequate SNOMED coverage. Table 3 demonstrates example which illustrate the limitations associated with the use of automatic coding tools for uncorrected text. Spelling errors such as fat igue not surprisingly result in inaccurate coding. Similarly many common colloquialisms such as labs instead of laboratory tests are not currently included in SNOMED CT and cannot be accurately coded. Unfortunately, a significant number of URN terms could simply not be interpreted at all as a result of the use of non-standard abbreviations, spelling errors, non-standard terms or a combination of these factors. For example, given the clinical source of the URN terms, Tuber most likely represents tuberculosis rather than a fleshy root or bulb. More interesting, though, are the terms which simply do not exist in SNOMED or, because of ambiguity in the language of the URN, are inaccurately coded. The example of tetanus is representative of the latter situation. In this case tetanus, which appeared numerous times in the list of 5054 URN s, is almost certainly is being entered as a short form 541

Table 2: Assessment of SNOMED codes generated by automatic tools and expert review with manual look up. The first rows demonstrate the accuracy of the tool outputs when reviewed by clinical experts. The unknown group included all URN s which remained ambiguous or uninterpretable after spelling correction and expansion of abbreviations. oftetanus toxoid (since frequency of occurrence of this URN is greatly disproportionate to the relative incidence of the disease in the US) but is apparently reasonably coded as the disease of tetanus. A total of 108 of the sampled URN s (11.9%) remained unresolved even with expert manual review and coding, largely due to due inability to determine the original meaning of the URN. This was generally related to the entry of ambiguous abbreviations, severe spelling errors (for example, vertiligo probably representing vertigo vs. vitiligo), or purely numeric data, which likely reflected ICD-9 or CPT codes. Post-coordination was required for many of the URN s for which manual review and coding resulted in accurate matches of the main atomic concepts. A much smaller number of URN s could be represented by both a single pre-coordinated term and by multiple atomic elements requiring post-coordination. We assessed for the presence of necessary linking semantic concepts in 117 URN s requiring post-coordination taken from the manually reviewed, randomized sample of 400 URN s. In this group 89% of appropriate semantic links were found. If this is extrapolated to the remainder URN s, it is estimated that, while 95.8 99% of major conceptual atoms can be found, only 85-89% of these have linking concepts to fully describe the entire, often complex, URN concept. For example, lower extremity weakness can be matched to entire lower limb (182281004) and asthenia (13791008) but a full association between these two atoms requires the addition of a linking concept such as finding site. Table 3. Examples of manually assessed accuracy for URN terms with identical SNOMED code output from both tools. Accuracy URN SNOMED concept ID SNOMED descriptor Partial Cecal Diverticulitis (broad) diverticulitis 307496006 Fasting labs 16985007 Fasting SP BTL 237679004 Status post Incorrect Fat igue 256674009 Fat Tetanus (vaccine) 76902006 Tetanus (disease) Unknown Hem 57813001 Heme EP 50762005 IL-1 Tuber 75505008 Tuber (root) Discussion It has been reported previously that common reasons for failure to match VA URN s with coded concepts in the VA Lexicon included use of numeric data or modifiers [7]. Similar types of problems also limited matching SNOMED codes in this study. It is also notable that a significant proportion of URN s remained unmatched even after expert review. Obviously, the ability to appropriately codify these types of data depends upon entry of accurate and unambiguous terms into the database, in addition to the existence of matching coded concepts. 542

Redundancy in SNOMED concepts was an interesting finding and was illustrated by the tool outputs. The TermWorks tool provided a single best pre-coordinated code while Opticode parsed each URN into more atomic elements. There were 104 of 303 URN s (greater than 30% of the failed match category) that remained unresolved even after expert review. Clearly the quality of the data is an important factor explaining match failures in this category. There were several instances where each tool provided different codes but taken together the codes gave conceptually identical results. For example, colon cancer (SNOMED-CT code 363406005) would be considered equivalent to adenocarcinoma (35917007) of the colon (302508007). In general, more complex the term, the more theoretical synonymous combinations will be possible. In addition the original language of the term will likely affect automated coding and possibly manual coding as well. A Billroth II operation could be considered verbal shorthand for an arguably more precise or specific antrectomy (or distal gastrectomy or partial gastrectomy) with retrocolic (or antecolic) gastojejunostomy. While an abdominal surgeon may understand the nuances of the elements contained in the latter term a clinician in a different specialty might not. As a result, each clinician may code such terms slightly differently, relative to their particular understanding of the term. The use of pre-coordinated rather than more atomic concepts leads to issues surrounding the benefits and hazards of combinatorial vocabularies [13]. Currently SNOMED allows both strategies and this may create additional complexity in extracting concepts for use in population based research, decision support and other clinical purposes. Additionally, the use of post-coordination requires the presence of appropriate semantic linking concepts and our results in a small sample show that 89% of these necessary concepts do exist. However, this also implies that in 11% of cases the entire meaning of the URN cannot be fully described by the main atomic elements without a linking concept that does not appear to exist in SNOMED at this time. Further study is necessary to determine if these rates are representative of other narrative sources and clinical domains. There are several limitations of this study. The existence of linking concepts was only evaluated in a small sample of URN s (117 terms requiring post-coordination) and further study will be needed to validate the results based on these data. Additionally, the high frequency of spelling errors and jargon in the unresolved narratives reduced the effectiveness of the automated matching tools and decreased our ability to assess their overall value. However, in data subsets having standard clinical language these tools appear promising. Conclusion This study demonstrates that greater than 85% of VHA s unresolved narratives can, in fact, be coded using SNOMED-CT. It is likely that more structured sources such as orders and operative or progress notes would have similar or perhaps even better coverage by SNOMED. This suggests that the addition of SNOMED-CT coding and semantic relationships to the VHA Health Data Repository will allow considerably greater utility than currently exists. Acknowledgements Drs. Lincoln, Penz, Nguyen and Sims gratefully acknowledge support from the Department of Veteran Affairs and the VHA Salt Lake City OI Field Office. The Clue-5 Browser was obtained by download from (http://www.clinical-info.co.uk/), The Clinical Information Consultancy, Reading Berkshire, UK. References [1] http://www.va.gov/vista_monograph (2003) [2] http://www.snomed.org, (2003). "SNOMED Clinical Terms (SNOMED CT ) Core Content as of January 2003." 1-2. [3] Campbell JR, Carpenter P, Sneiderman C, Cohn S, Chute CG, Warren J. (1997). "Phase II evaluation of clinical coding schemes: completeness, taxonomy, mapping, definitions, and clarity. CPRI Work Group on Codes and Structures." J Am Med Inform Assoc. 1997 May-Jun; 4(3): 254-5. [4] Chute CG, Cohn SP, Campbell KE, Oliver DE, Campbell JR. (1996). "The content coverage of clinical classifications. For The Computer-Based Patient Record Institute's Work Group on Codes & Structures." J Am Med Inform Assoc. 3(3): 224 233. [5] Chute CG, Cohn SP, Campbell JR. (1998). "A Framework for Comprehensive Health Terminology Systems in the United States Development Guidelines, Criteria for Selection, and Public Policy Implications." J Am Med Inform Assoc. 5(6): 503 510. [6] Cimino, JJ. (1998) Desiderata for Controlled Medical Vocabularies in the Twenty-First Century. Methods of Information in Medicine; 37:394-403. [7] Zinder DJ. Structured documentation. Otolaryngologic Clinics of North America. 2002 Dec; 35(6): 1211-21. [8] Eagon JC, Hurdle JF, Lincoln MJ (1996). "Inter-rater reliability and review of the VA unresolved narratives." Proc AMIA Annu Fall Symp: 130-4. [9] K J Carter, S. R., E Kessler, L P Caccamo, N P Ritchey, B A Erickson, F Castro, and M D Poggione (1996). "Quality assurance in anatomic pathology: automated SNOMED coding." J Am Med Inform Assoc. 3(4): 270 272. [10]Kamel Boulos MN, Roudsari AV, Carson ER. (2002). "A dynamic problem to knowledge linking Semantic Web service based on clinical codes." Med Inform Internet Med 27(3): 127-37. [11]Dolin RH, Huff SM, Rocha RA, Spackman KA, Campbell KE (1998). "Evaluation of a Lexically Assign, Logically Refine Strategy for Semi-automated Integration of Overlapping Terminologies." J Am Med Inform Assoc. 5(2): 203 213. [12]Brown SH, Lincoln MJ, Groen PJ, Kolodner RM VistA-- U.S. Department of Veterans Affairs national-scale health information system. Int J Med Inf. 2003 Mar; 69(2-3):135-56. 543

[13]Elkin PL, Brown SH, Chute CG. Guideline for Health Informatics: Controlled Health Vocabularies Vocabulary Structure and High-Level Indicators. MedInfo (2001); 10 (part 1), 191-5. Address for Correspondence Janet FE Penz, MD, Department of Surgery, Veterans Administration Medical Center, 500 Foothill Drive, Salt Lake City, UT 84148 544