SNUMedinfo at TREC CDS track 2014: Medical case-based retrieval task

Similar documents
Exploring the Query Expansion Methods for Concept Based Representation

Classification and retrieval of biomedical literatures: SNUMedinfo at CLEF QA track BioASQ 2014

Northeastern University in TREC 2009 Million Query Track

BUPT at TREC 2009: Entity Track

Using Model-Theoretic Invariants for Semantic Integration. Michael Gruninger NIST / Institute for Systems Research University of Maryland

Multi-Modal Communication

UMass at TREC 2006: Enterprise Track

Better Contextual Suggestions in ClueWeb12 Using Domain Knowledge Inferred from The Open Web

Service Level Agreements: An Approach to Software Lifecycle Management. CDR Leonard Gaines Naval Supply Systems Command 29 January 2003

FUDSChem. Brian Jordan With the assistance of Deb Walker. Formerly Used Defense Site Chemistry Database. USACE-Albuquerque District.

4. Lessons Learned in Introducing MBSE: 2009 to 2012

ICTNET at Web Track 2010 Diversity Task

PRIS at TREC2012 KBA Track

DoD Common Access Card Information Brief. Smart Card Project Managers Group

Kathleen Fisher Program Manager, Information Innovation Office

Empirically Based Analysis: The DDoS Case

COMPUTATIONAL FLUID DYNAMICS (CFD) ANALYSIS AND DEVELOPMENT OF HALON- REPLACEMENT FIRE EXTINGUISHING SYSTEMS (PHASE II)

Experiments with ClueWeb09: Relevance Feedback and Web Tracks

Introducing I 3 CON. The Information Interpretation and Integration Conference

Dana Sinno MIT Lincoln Laboratory 244 Wood Street Lexington, MA phone:

Experiments on Related Entity Finding Track at TREC 2009 Qing Yang,Peng Jiang, Chunxia Zhang, Zhendong Niu

BIT at TREC 2009 Faceted Blog Distillation Task

A Review of the 2007 Air Force Inaugural Sustainability Report

Running CyberCIEGE on Linux without Windows

2013 US State of Cybercrime Survey

Information, Decision, & Complex Networks AFOSR/RTC Overview

ASSESSMENT OF A BAYESIAN MODEL AND TEST VALIDATION METHOD

Full-texts representation with Medical Subject Headings, and co-citations network reranking strategies for TREC 2014 Clinical Decision Support Track

ARINC653 AADL Annex Update

Vision Protection Army Technology Objective (ATO) Overview for GVSET VIP Day. Sensors from Laser Weapons Date: 17 Jul 09 UNCLASSIFIED

MODELING AND SIMULATION OF LIQUID MOLDING PROCESSES. Pavel Simacek Center for Composite Materials University of Delaware

Architecting for Resiliency Army s Common Operating Environment (COE) SERC

University of TREC 2009: Indexing half a billion web pages

University of Waterloo: Logistic Regression and Reciprocal Rank Fusion at the Microblog Track

Web Site update. 21st HCAT Program Review Toronto, September 26, Keith Legg

University of Pittsburgh at TREC 2014 Microblog Track

Using Templates to Support Crisis Action Mission Planning

ENVIRONMENTAL MANAGEMENT SYSTEM WEB SITE (EMSWeb)

Space and Missile Systems Center

High-Assurance Security/Safety on HPEC Systems: an Oxymoron?

Concept of Operations Discussion Summary

A Distributed Parallel Processing System for Command and Control Imagery

Topology Control from Bottom to Top

Setting the Standard for Real-Time Digital Signal Processing Pentek Seminar Series. Digital IF Standardization

A Methodology for End-to-End Evaluation of Arabic Document Image Processing Software

QCRI at TREC 2014: Applying the KISS principle for the TTG task in the Microblog Track

The C2 Workstation and Data Replication over Disadvantaged Tactical Communication Links

75th Air Base Wing. Effective Data Stewarding Measures in Support of EESOH-MIS

Data Warehousing. HCAT Program Review Park City, UT July Keith Legg,

Accuracy of Computed Water Surface Profiles

Opinions in Federated Search: University of Lugano at TREC 2014 Federated Web Search Track

Dr. Stuart Dickinson Dr. Donald H. Steinbrecher Naval Undersea Warfare Center, Newport, RI May 10, 2011

Corrosion Prevention and Control Database. Bob Barbin 07 February 2011 ASETSDefense 2011

On Duplicate Results in a Search Session

Towards a Formal Pedigree Ontology for Level-One Sensor Fusion

Lessons Learned in Adapting a Software System to a Micro Computer

DEVELOPMENT OF A NOVEL MICROWAVE RADAR SYSTEM USING ANGULAR CORRELATION FOR THE DETECTION OF BURIED OBJECTS IN SANDY SOILS

CENTER FOR ADVANCED ENERGY SYSTEM Rutgers University. Field Management for Industrial Assessment Centers Appointed By USDOE

73rd MORSS CD Cover Page UNCLASSIFIED DISCLOSURE FORM CD Presentation

David W. Hyde US Army Engineer Waterways Experiment Station Vicksburg, Mississippi ABSTRACT

Distributed Real-Time Embedded Video Processing

Advanced Numerical Methods for Numerical Weather Prediction

Space and Missile Systems Center

FUB, IASI-CNR and University of Tor Vergata at TREC 2008 Blog Track

Technological Advances In Emergency Management

Shallow Ocean Bottom BRDF Prediction, Modeling, and Inversion via Simulation with Surface/Volume Data Derived from X-Ray Tomography

Cyber Threat Prioritization

ATCCIS Replication Mechanism (ARM)

Defense Hotline Allegations Concerning Contractor-Invoiced Travel for U.S. Army Corps of Engineers' Contracts W912DY-10-D-0014 and W912DY-10-D-0024

Washington University

Microsoft Research Asia at the Web Track of TREC 2009

C2-Simulation Interoperability in NATO

73rd MORSS CD Cover Page UNCLASSIFIED DISCLOSURE FORM CD Presentation

Use of the Polarized Radiance Distribution Camera System in the RADYO Program

Ross Lazarus MB,BS MPH

Advanced Numerical Methods for Numerical Weather Prediction

A New Approach to Contextual Suggestions Based on Word2Vec

Wireless Connectivity of Swarms in Presence of Obstacles

2011 NNI Environment, Health, and Safety Research Strategy

SURVIVABILITY ENHANCED RUN-FLAT

Col Jaap de Die Chairman Steering Committee CEPA-11. NMSG, October

ASPECTS OF USE OF CFD FOR UAV CONFIGURATION DESIGN

INTEGRATING LOCAL AND GLOBAL NAVIGATION IN UNMANNED GROUND VEHICLES

Office of Global Maritime Situational Awareness

U.S. Army Research, Development and Engineering Command (IDAS) Briefer: Jason Morse ARMED Team Leader Ground System Survivability, TARDEC

Basing a Modeling Environment on a General Purpose Theorem Prover

LARGE AREA, REAL TIME INSPECTION OF ROCKET MOTORS USING A NOVEL HANDHELD ULTRASOUND CAMERA

Detection, Classification, & Identification of Objects in Cluttered Images

Speaker Verification Using SVM

Energy Security: A Global Challenge

DoD M&S Project: Standardized Documentation for Verification, Validation, and Accreditation

QuanTM Architecture for Web Services

WAITING ON MORE THAN 64 HANDLES

CNIPA, FUB and University of Rome Tor Vergata at TREC 2008 Legal Track

Computer Aided Munitions Storage Planning

An Efficient Architecture for Ultra Long FFTs in FPGAs and ASICs

Approaches to Improving Transmon Qubits

Automation Middleware and Algorithms for Robotic Underwater Sensor Networks

AIS Indexer User Guide

MATREX Run Time Interface (RTI) DoD M&S Conference 10 March 2008

Transcription:

SNUMedinfo at TREC CDS track 2014: Medical case-based retrieval task Sungbin Choi, Jinwook Choi Medical Informatics Laboratory, Seoul National University, Seoul, Republic of Korea wakeup06@empas.com, jinchoi@snu.ac.kr Abstract. This paper describes the participation of the SNUMedinfo team at the TREC Clinical Decision Support track 2014. This task is about medical casebased retrieval. Case description is used as query text. Per each query, one of three categories (Diagnosis, Test and Treatment) is designated as target information need. Firstly, we used external tagged knowledge-based query expansion method for the relevance ranking. Secondly, machine learning classifier based text categorization method is used for the task-specific ranking. Finally, we combined relevance ranking and task-specific ranking with method. Our method showed significant performance improvements. Keywords: Case-based retrieval, Query expansion, Text categorization, Information retrieval 1. Introduction In this paper, we describe the methods in participation of the SNUMedinfo team at the TREC Clinical Decision Support (CDS) track 2014. The task is about medical case-based retrieval task. Case description is used as query text. Per each query, one of three category (Diagnosis, Test and Treatment) is designated as target information need. For detailed task introduction, please see the overview paper of this track. 2. Methods Our method can be summarized as following three steps (Section 2.1 to 2.3)

Report Documentation Page Form Approved OMB No. 0704-0188 Public reporting burden for the collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing this burden, to Washington Headquarters Services, Directorate for Information Operations and Reports, 1215 Jefferson Davis Highway, Suite 1204, Arlington VA 22202-4302. Respondents should be aware that notwithstanding any other provision of law, no person shall be subject to a penalty for failing to comply with a collection of information if it does not display a currently valid OMB control number. 1. REPORT DATE NOV 2014 2. REPORT TYPE 3. DATES COVERED 00-00-2014 to 00-00-2014 4. TITLE AND SUBTITLE SNUMedinfo at TREC CDS Track 2014: Medical Case-based Retrieval Task 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) 5d. PROJECT NUMBER 5e. TASK NUMBER 5f. WORK UNIT NUMBER 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) Seoul National University,Medical Informatics Laboratory,Seoul, Republic of Korea, 8. PERFORMING ORGANIZATION REPORT NUMBER 9. SPONSORING/MONITORING AGENCY NAME(S) AND ADDRESS(ES) 10. SPONSOR/MONITOR S ACRONYM(S) 12. DISTRIBUTION/AVAILABILITY STATEMENT Approved for public release; distribution unlimited 11. SPONSOR/MONITOR S REPORT NUMBER(S) 13. SUPPLEMENTARY NOTES presented in the proceedings of the Twenty-Third Text REtrieval Conference (TREC 2014) held in Gaithersburg, Maryland, November 19-21, 2014. The conference was co-sponsored by the National Institute of Standards and Technology (NIST) and the Defense Advanced Research Projects Agency (DARPA). 14. ABSTRACT This paper describes the participation of the SNUMedinfo team at the TREC Clinical Decision Support track 2014. This task is about medical case-based retrieval. Case description is used as query text. Per each query, one of three categories (Diagnosis, Test and Treatment) is designated as target infor-mation need. Firstly, we used external tagged knowledge-based query expan-sion method for the relevance ranking. Secondly, machine learning classifier based text categorization method is used for the task-specific ranking. Finally, we combined relevance ranking and task-specific ranking with method. Our method showed significant performance improvements. 15. SUBJECT TERMS 16. SECURITY CLASSIFICATION OF: 17. LIMITATION OF ABSTRACT a. REPORT unclassified b. ABSTRACT unclassified c. THIS PAGE unclassified Same as Report (SAR) 18. NUMBER OF PAGES 7 19a. NAME OF RESPONSIBLE PERSON Standard Form 298 (Rev. 8-98) Prescribed by ANSI Std Z39-18

2.1 External tagged knowledge-based query expansion We used external medical literature corpus (MEDLINE ) as a tagged knowledge source to acquire useful query expansion terms. We leased the 2014 MEDLINE /PubMed Journal Citations from the U.S. National Library of Medicine. There are approximately 22 million MEDLINE citations. Article title, abstract text, MeSH descriptor fields are indexed. We used the unigram query likelihood (QL) model [1] with Dirichlet prior smoothing [2] as our baseline retrieval model. The Indri search engine [3] was used in the experiment. The queries are stopped at the query time using the standard 418 INQUERY stopword list, case-folded, and stemmed using Porter stemmer. Per each original case query, we retrieved relevant documents from external corpus (MEDLINE) using query likelihood model. We extracted MeSH MajorTopic descriptors from top-k ranked documents. The original case query is expanded with these MeSH MajorTopic terms. Using this expanded query, we retrieved 1,000 documents per each query from target corpus (TREC CDS track). The Indri query is described as follows. #weight ( (1-w) #combine (original query terms) w #combine (expansion query terms) ) Similar method showed effective performance in our previous study 1 [4] (ImageCLEF case-based retrieval task 2013 [5]). 2.2 Task-specific ranking Per each query, one of three category (Diagnosis, Test and Treatment) is designated as target information need. We trained task classifiers on the Clinical Hedges database [6] and applied them on the top 1,000 documents from Section 2.1 to have taskspecific ranking. In Clinical Hedges database, documents are manually classified by purpose category (e.g., therapy, diagnosis, prognosis). We trained two task classifiers; CHD_TR_Classifier is trained to classify therapy versus non- therapy documents. CHD_DX_Classifier is trained to classify diagnosis versus non- diagnosis documents. SVM-perf [7] is used for the classification task. Both classifiers are trained to optimize AUC (area under the ROC curve). Trained classifiers are applied on the top 1,000 documents from Section 2.1. Then, documents are sorted by classification score. 2.3 Combining relevance ranking with task-specific ranking We combined relevance ranking and task-specific ranking with method [8]. When different aspects need to be considered together for the document ranking, method showed effective performance in our previous experiment [9]. 1 Compared to our method used in the ImageCLEF 2013 case-based retrieval task, this time we didn t apply limitation on the publication type of pseudo-relevant documents. We found out that it is not helpful to improve performance in our additional experiments on the ImageCLEF 2013 test set.

2.4 Submitted runs Details of our submitted runs can be summarized as following table. Table 1. Submitted runs RunID Query Version Details per Query type SNUMedinfo1 Summary Diagnosis : SNUMedinfo2 Summary Diagnosis : (Relevance ranking + rank_min 2 (CHD_DX_Classifier, CHD_TR_Classifier) ) SNUMedinfo3 Summary Diagnosis : Relevance ranking only 2 For example, if Document A is ranked 10 th by CHD_DX_Classifier, and ranked 800 th by CHD_TR_Classifier, then output of rank_min for Document A is 10. If Document B is ranked 900 th by CHD_DX_Classifier, and ranked 100 th by CHD_TR_Classifier, then output of rank_min for Document B is 100.

SNUMedinfo4 Description Diagnosis : SNUMedinfo5 (No submit) Description Diagnosis : (Relevance ranking + rank_min(chd_dx_classifier, CHD_TR_Classifier) ) SNUMedinfo6 Description Diagnosis : Relevance ranking only Query type Treatment is considered to be match with CHD_TR_Classifier. Query type Test is considered to be match with CHD_DX_Classifier. With regard to the query type Diagnosis, by definition it is considered equivalent to the query types used in ImageCLEF case-based retrieval task [5], and that s why we applied only relevance ranking in SNUMedinfo3, SNUMedinfo6. But on the other hand, we thought that also it could be helpful to combine other task-specific ranking with relevance ranking, because Test or Treatment tasks are closely related to the Diagnosis task.

3. Results Table 2. Evaluation results (query version: Summary) RunID infndcg infap P@10 Baseline (QL) 0.1921 0.0501 0.3400 ExternalQE 0.2224 0.0589 0.3200 SNUMedinfo1 0.2188 0.0463 0.3367 SNUMedinfo2 0.2173 0.0458 0.3333 SNUMedinfo3 0.2406 0.0582 0.3467 QL : Query likelihood model with original query ExternalQE : External tagged knowledge based query expansion Best result per column is marked in boldface Table 3. Evaluation results (query version: Description) RunID infndcg infap P@10 Baseline (QL) 0.1877 0.0436 0.2933 ExternalQE 0.2199 0.0511 0.3200 SNUMedinfo4 0.2502 0.0545 0.3300 SNUMedinfo5 0.2505 0.0556 0.3267 SNUMedinfo6 0.2674 0.0659 0.3633 QL : Query likelihood model with original query ExternalQE : External tagged knowledge based query expansion Best result per column is marked in boldface In Table 2, SNUMedinfo3 showed significant performance improvement over baseline. In Table 3, SNUMedinfo6 showed significant performance improvement over baseline. Both SNUMedinfo3 and SNUMedinfo6 used relevance ranking only for the query type Diagnosis, while of relevance ranking and task-specific ranking is used for the Test and Treatment query type. 4. Discussion In Table 4 and Table 5, we compared evaluation results of different methods per query type. Table 4. Comparison of results per query type (query version: Summary, evaluation metric : infndcg) Diagnosis Test Treatment Total Baseline 0.2263 0.1515 0.1984 0.1921 ExternalQE 0.2945 0.1546 0.2182 0.2224 SNUMedinfo3 0.2945 0.1831 0.2443 0.2406 ExternalQE : External tagged knowledge based query expansion

Table 5. Comparison of results per query type (query version: Description, evaluation metric : infndcg) Diagnosis Test Treatment Total Baseline 0.2270 0.1558 0.1804 0.1877 ExternalQE 0.2977 0.1346 0.2273 0.2199 SNUMedinfo6 0.2977 0.2029 0.3016 0.2674 ExternalQE : External tagged knowledge based query expansion 5. Conclusion TREC CDS 2014 was a medical case-based retrieval task, and each query had different target task among diagnosis, test or treatment. As a first step, we used external tagged knowledge based query expansion method to retrieve relevant documents. As a second step, we trained machine learning document classifier to compute taskspecific ranking of documents. Finally, we combined relevance ranking and taskspecific ranking with method. Our method showed significant improvement over baseline method. 6. Acknowledgements Use of the Clinical Hedges database was made possible through a collaboration agreement with R B Haynes and N L Wilczynski at McMaster University, Hamilton, Ontario Canada. This study was supported by a grant of the Korean Health Technology R&D Project, Ministry of Health & Welfare, Republic of Korea. (No. HI11C1947) 7. References 1. Ponte, J.M. and W.B. Croft, A language modeling approach to information retrieval, in Proceedings of the 21st annual international ACM SIGIR conference on Research and development in information retrieval. 1998, ACM: Melbourne, Australia. p. 275-281. 2. Zhai, C. and J. Lafferty, A study of smoothing methods for language models applied to Ad Hoc information retrieval, in Proceedings of the 24th annual international ACM SIGIR conference on Research and development in information retrieval. 2001, ACM: New Orleans, Louisiana, USA. p. 334-342. 3. Strohman, T., et al. Indri: A language model-based search engine for complex queries. in Proceedings of the International Conference on Intelligent Analysis. 2005. McLean, VA.

4. Choi, S., J. Lee, and J. Choi. SNUMedinfo at ImageCLEF 2013: Medical retrieval task. in CLEF (Online Working Notes/Labs/Workshop). 2013. 5. Garcıa Seco de Herrera, A., et al. Overview of the ImageCLEF 2013 medical tasks. Working Notes for CLEF 2013 Conference, 2013. 1179. 6. Haynes, R.B., et al., Optimal search strategies for retrieving scientifically strong studies of treatment from Medline: analytical survey. Vol. 330. 2005. 1179. 7. Joachims, T., Training linear SVMs in linear time, in Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining. 2006, ACM: Philadelphia, PA, USA. p. 217-226. 8. Aslam, J.A. and M. Montague, Models for metasearch, in Proceedings of the 24th annual international ACM SIGIR conference on Research and development in information retrieval. 2001, ACM: New Orleans, Louisiana, USA. p. 276-284. 9. Choi, S., et al., Combining relevancy and methodological quality into a single ranking for evidence-based medicine. Information Sciences, 2012. 214(0): p. 76-90.