Entity Linking at TAC Task Description

Similar documents
Wikipedia and the Web of Confusable Entities: Experience from Entity Linking Query Creation for TAC 2009 Knowledge Base Population

2017 TAC KBP Event Detection and Coreference Tasks

UBC Entity Discovery and Linking & Diagnostic Entity Linking at TAC-KBP 2014

NUS-I2R: Learning a Combined System for Entity Linking

Linguistic Resources for Entity Linking Evaluation: from Monolingual to Cross-lingual

Proposed Task Description for Knowledge-Base Population at TAC English Slot Filling Regular and Temporal

Graph-Based Entity Linking

Linguistic Resources for 2013 Knowledge Base Population Evaluations

PRIS at TAC2012 KBP Track

Entity Linking. David Soares Batista. November 11, Disciplina de Recuperação de Informação, Instituto Superior Técnico

CMU System for Entity Discovery and Linking at TAC-KBP 2017

CMU System for Entity Discovery and Linking at TAC-KBP 2016

Annotating Spatio-Temporal Information in Documents

Stacked Ensembles of Information Extractors for Knowledge-Base Population

CMU System for Entity Discovery and Linking at TAC-KBP 2015

Bi-directional Linkability From Wikipedia to Documents and Back Again: UMass at TREC 2012 Knowledge Base Acceleration Track

ACE 2008: Cross-Document Annotation Guidelines (XDOC)

Proposed Task Description for Source/Target Belief and Sentiment Evaluation (BeSt) at TAC 2016

Stanford-UBC at TAC-KBP

New York University 2014 Knowledge Base Population Systems

Document Structure Analysis in Associative Patent Retrieval

The Wikipedia XML Corpus

The ALBAYZIN 2016 Search on Speech Evaluation Plan

Benchmarks for Enterprise Linking: Thomson Reuters R&D at TAC 2013

Dublin City University at CLEF 2005: Multi-8 Two-Years-On Merging Experiments

Ghent University-IBCN Participation in TAC-KBP 2015 Cold Start Slot Filling task

Better Evaluation for Grammatical Error Correction

Stanford s 2013 KBP System

WebTLab: A cooccurrence-based approach to KBP 2010 Entity-Linking task

Towards Summarizing the Web of Entities

A Comparative Study Weighting Schemes for Double Scoring Technique

Assignment 3 ITCS-6010/8010: Cloud Computing for Data Analysis

HUKB at NTCIR-12 IMine-2 task: Utilization of Query Analysis Results and Wikipedia Data for Subtopic Mining

ABSTRACT. Professor Douglas W. Oard College of Information Studies

Tulip: Lightweight Entity Recognition and Disambiguation Using Wikipedia-Based Topic Centroids. Marek Lipczak Arash Koushkestani Evangelos Milios

Hyperlink-Induced Topic Search (HITS) over Wikipedia Articles using Apache Spark

Proceedings of NTCIR-9 Workshop Meeting, December 6-9, 2011, Tokyo, Japan

A simple method to extract abbreviations within a document using regular expressions

Overview of the INEX 2009 Link the Wiki Track

Today s topic CS347. Results list clustering example. Why cluster documents. Clustering documents. Lecture 8 May 7, 2001 Prabhakar Raghavan

IRCE at the NTCIR-12 IMine-2 Task

Overview of TAC-KBP 2015 Event Nugget Track

Ranked Retrieval. Evaluation in IR. One option is to average the precision scores at discrete. points on the ROC curve But which points?

MIRACLE at ImageCLEFmed 2008: Evaluating Strategies for Automatic Topic Expansion

Towards Breaking the Quality Curse. AWebQuerying Web-Querying Approach to Web People Search.

Dialog System & Technology Challenge 6 Overview of Track 1 - End-to-End Goal-Oriented Dialog learning

Mining Wikipedia s Snippets Graph: First Step to Build A New Knowledge Base

SAS BI Dashboard 3.1. User s Guide Second Edition

Using an Image-Text Parallel Corpus and the Web for Query Expansion in Cross-Language Image Retrieval

Understanding the Query: THCIB and THUIS at NTCIR-10 Intent Task. Junjun Wang 2013/4/22

Lecture 8 May 7, Prabhakar Raghavan

How Co-Occurrence can Complement Semantics?

DCU at FIRE 2013: Cross-Language!ndian News Story Search

Cost-Effective Conceptual Design. over Taxonomies. Yodsawalai Chodpathumwan. University of Illinois at Urbana-Champaign.

CSCI 320 Group Project

Research Tools: DIY Text Tools

Chinese Question Answering using the DLT System at NTCIR 2005

Chinese Microblog Entity Linking System Combining Wikipedia and Search Engine Retrieval Results

Stanford-UBC Entity Linking at TAC-KBP

Wikipedia Retrieval Task ImageCLEF 2011

Precise Medication Extraction using Agile Text Mining

CS47300: Web Information Search and Management

Plenipotentiary Conference (PP- 14) Busan, 20 October 7 November 2014

Temporal Information Access (Temporalia-2)

Event Registry. Gregor Leban. Jozef Stefan Institute

Synology Security Whitepaper

MT in the Online Environment: Challenges and Opportunities

Mastering the Grade Book

ResPubliQA 2010

Towards open-domain QA. Question answering. TReC QA framework. TReC QA: evaluation

Overview of the 2013 ALTA Shared Task

Retrieval Evaluation. Hongning Wang

SYDNEY CMCRC at TAC

Plenipotentiary Conference (PP- 14) Busan, 20 October 7 November 2014

Data Preprocessing. Slides by: Shree Jaswal

From CLIR to CLIE: Some Lessons in NTCIR Evaluation

IBM Watson Application Developer Workshop. Watson Knowledge Studio: Building a Machine-learning Annotator with Watson Knowledge Studio.

Implementation of a High-Performance Distributed Web Crawler and Big Data Applications with Husky

CIRGDISCO at RepLab2012 Filtering Task: A Two-Pass Approach for Company Name Disambiguation in Tweets

WELCOME TO A SPANISH SPEAKING WORLD

Yahoo! Webscope Datasets Catalog January 2009, 19 Datasets Available

High Performance Oracle Endeca Designs for Retail. Technical White Paper 24 June

Understanding the Query: THCIB and THUIS at NTCIR-10 Intent Task

Data Mining Concepts & Techniques

Better translations with user collaboration - Integrated MT at Microsoft

Semantic Search at Bloomberg

A cocktail approach to the VideoCLEF 09 linking task

PRIS at TREC2012 KBA Track

Production Report of the TRAD Parallel Corpus Chinese-French Translation of a subset of GALE Phase 1 Chinese Blog Parallel Text (LDC2008T06)

Module 3: GATE and Social Media. Part 4. Named entities

Challenges on Combining Open Web and Dataset Evaluation Results: The Case of the Contextual Suggestion Track

This corpus has been produced within the framework of the PEA-TRAD project ( )

User guide for GEM-TREND

Information Retrieval

Information Retrieval. Lecture 7 - Evaluation in Information Retrieval. Introduction. Overview. Standard test collection. Wintersemester 2007

Information Retrieval and Organisation

Course on Database Design Carlo Batini University of Milano Bicocca Part 5 Logical Design

Data Dictionary. National Center for O*NET Development

Developing Focused Crawlers for Genre Specific Search Engines

Creating N-gram profile for a Wikipedia Corpus

Transcription:

Entity Linking at TAC 2013 Task Description Version 1.0 of April 9, 2013 1 Introduction The main goal of the Knowledge Base Population (KBP) track at TAC 2013 is to promote research in and to evaluate the ability of automated systems to discover information about named entities and to incorporate this information in a knowledge source. For the evaluation an initial (or reference) knowledge base will be provided along with a source document collection from which systems are to learn. Attributes (a.k.a., slots ) derived from Wikipedia Infoboxes will be used to create the reference knowledge base. The overall task of populating a knowledge base is decomposed into two related tasks: Entity Linking (EL), where names must be aligned to entities; and Slot Filling (SF), which involves mining information about entities from text. In the Entity Linking (EL) task, names must be aligned to entities in the reference KB and with other entities discovered in the collection. The task will be structured by having participants process a list of queries over target entities. This list will contain entity types of Person (PER), Organization (ORG), and Geo-Political Entity (GPE). As in the ACE evaluation, GPEs include inhabited locations with a government, such as cities and countries. 2 Entity Linking 2.1 Monolingual Entity Linking In the Entity Linking task, given a query that consists of a name string, a background document ID, and a pair of UTF-8 character offsets indicating the beginning and end locations of the name string in the document, the system is required to provide the ID of the KB entry to which the name refers, or a NILxxxx ID if there is no such KB entry. The entity linking system is required to cluster queries referring to the same non-kb (NIL) entities and provide a unique ID for each cluster, in the form of NILxxxx (e.g., NIL0021 ). An example query is <query id="el_000304"> <name>robinson</name> <docid>apw_eng_20080416.1425.ldc2009t13</docid> <beg>565</beg> <end>572</end> </query> The entity-linking system output file should contain one line for each query, where each line has three tab-delimited fields: Field 1: query ID Field 2: reference KB link (or NIL link) Field 3: a confidence value Each confidence value must be a positive real number between 0.0 (exclusive, representing the lowest confidence) and 1.0 (inclusive, representing the highest confidence), and must 1

include a decimal point (no commas, please). Up to five answers to a given query may be included in each submission. The main score for the task will use only the highest confidence answer for each query, selecting the answer that appears earliest in the submission if more than one answer has the highest confidence value. If practical given resource limitations, NIST will provide alternate scoring that looks at all submitted answers. At this point it is not feasible to allow multiple NIL clusterings; thus, no more than one NIL value may be submitted for a given query, and NIL clustering is not likely to play a part in the alternate scoring. Entities will generally occur in multiple queries using different name variants and/or different docids. It is also expected that some entities will share confusable names (e.g., George Washington could refer to the president, the university, or the jazz musician; Washington could refer to a city, state, or person). For the primary task, the system may consult the text from the Wikipedia pages associated with the KB nodes. There will be also an optional task in which the systems should do linking without reference to these texts using only the slot values; this corresponds to the task of updating a knowledge base with no backing text. 2.2 Cross-lingual Entity Linking The cross-lingual entity linking tasks follow monolingual entity linking; the steps are: (1) link non-nil queries to English KB entries; and (2) cluster NIL queries. The cross-lingual aspect comes from the fact that the queries will include Chinese and Spanish queries. An example Chinese query is <query id="el_cmn_00007"> <name> 宜宾市 </name> <docid>pda_cmn_20070327.0068</docid> <beg>176</beg> <end>178</end> </query> 2.3 Scoring Metric For a set of query names with source documents, an entity linking system is required to: (1) judge whether each query can be linked to any KB node; and (2) partition all queries with NIL KB entries into clusters. Ultimately the system output can be viewed as a collection of clusters, some of which are labeled with KB node IDs. At the same time the answer key can also be viewed as a different collection of clusters over the same set. Therefore we will apply a modified B-Cubed metric (called B-Cubed+) to evaluate these clusters. Let us use the following notation: Le () and Ce () the category and the cluster of an item e, SI() e and GI() e represent, respectively, the system (i.e., participant-submitted) and goldstandard (ground truth) KB identifiers for an item e. We define the correctness of the relation between eand e ' in the distribution as: 1 iff L( e) = L( e') C( e) = C( e') GI( e) = SI( e) = GI( e') = SI( e') Gee (, ') = 0otherwise That is, two items are correctly related when they share a category if and only if they appear in the same cluster and share the same KB identifier in the system and the gold standard. B-cubed+ precision of an item is the proportion of correctly related items in its cluster (including itself). The 2

overall B-Cubed+ precision is the averaged precision of all items in the distribution. Since the average is calculated over items, it is not necessary to apply any weighting according to the size of clusters or categories. The B-Cubed+ recall is analogous, replacing cluster with category. Formally: Precision B-Cubed+ = Avg e [Avg e'.c (e)=c (e') [G(e,e')]] Recall B-Cubed+ = Avg e [Avg e'.l(e)=l(e') [G(e,e')]] The scorer will be available at: http://www.nist.gov/tac/2013/kbp/entitylinking/tools.html 3 Data 3.1 Knowledge Base and Source Document Collection The reference knowledge base includes hundreds of thousands of entities based on articles from an October 2008 dump of English Wikipedia, which includes 818,741 nodes. Each entity in the KB includes the following: a name string; an assigned entity type of PER, ORG, GPE, or UKN (unknown); 1 a KB node ID (a unique identifier, like E101 ); a set of raw (Wikipedia) slot names and values; and some disambiguating text (i.e., text from the Wikipedia page). The raw slot names and the values in the reference KB are based on the October 2008 Wikipedia snapshot. To facilitate use of the reference KB, a mapping from raw Wikipedia infobox slot-names to generic slots is provided in the training corpora. The source document collection for the KBP 2013 Entity Linking tasks are composed of English, Spanish, and Chinese documents from the following LDC packages: 1. LDC2011T13: Chinese Gigaword Fifth Edition 2. LDC2011T07: English Gigaword Fifth Edition 3. LDC2011T12: Spanish Gigaword Third Edition 4. LDC2012E23: TAC 2012 KBP Source Corpus Additions Web Documents 5. Approximately half a million discussion forum posts in English, Chinese, and possibly Spanish (TBA) Only a subset of the English, Spanish, and Chinese Gigaword collections will included in the official KBP 2013 tasks; the document IDs of those Gigaword documents that are part of the KBP 2013 tasks are listed in LDC catalog item LDC2012E22 (TAC 2012 KBP Source Corpus Additions Doc-ID Lists). The following Table 1 presents the profile of the collection of source documents for the KBP 2013 entity-linking tasks. Language Source Genre Size (documents) English LDC2011T07 1,000,257 (list in LDC2012E22) LDC2012E23 Web Text 999,999 TBA Discussion Fora ~500,000 Chinese LDC2011T13 2,000,256 (list in LDC2012E22) 1 The reason this is abbreviated UKN rather than UNK is UKN. 3

LDC2012E23 Web Text 815,886 TBA Discussion Fora ~500,000 Spanish LDC2011T12 1,000,020 (list in LDC2012E22) TBA Discussion Fora Up to ~500,000 Table 1. Distribution of Documents in KBP 2013 Source Document Collection 3.2 Training and Evaluation Corpus The following tables summarize the KBP 2013 training and evaluation data that we aim to provide for participants. For all tasks we try to achieve a balance among genres, and between the queries with and without KB entry linkages. Corpus Genre/Source Size (entity mentions) Person Organization GPE Training 2009 Eval 627 2710 567 2010 Training 500 500 500 Web data 2010 Eval 500 500 500 2010 Eval Web 250 250 250 data 2011 Eval 500 491 500 2011 Eval Web 250 259 250 data 2012 Eval 702 388 381 2012 Eval Web 216 318 221 Evaluation (estimate) data 2013 333 333 333 2013 333 333 333 Web/Discussion Fora data Table 2. English Monolingual Entity Linking Data Corpus Genre/Source Size (entity mentions) Person Organization GPE 2011 Training 250 250 250 Training English/Chinese 2011 Eval 250 250 250 English/Chinese 2012 Training Chinese Web 52 54 52 4

2012 Eval Chinese 466 442 454 2012 Eval Chinese Web 233 276 251 Evaluation (estimate) 2013 Eval Chinese 2013 Eval Chinese Web/Discussion Fora 333 333 333 333 333 333 Table 3. Chinese Cross-lingual Entity Linking Data Corpus Genre/Source Size (entity mentions) Person Organization GPE Training 2012 Training 664 597 563 Spanish 2012 Eval 669 539 858 Spanish Evaluation (estimate) 2013 Eval Spanish / Discussion Fora 667 667 667 Table 4. Spanish Cross-lingual Entity Linking Data 4 External Resource Restrictions and Sharing 4.1 External Resource Restrictions As in KBP 2012, participants will be asked to make at least one run subject to certain resource constraints, primarily that the run be made as a closed system one which does not access the Web during the evaluation period. Sites may also submit additional runs that access the Web. This will provide a better understanding of the impact of external resources. Further rules for both of the primary runs and additional runs are listed in Table 5. Specific Rules Allowed Not Allowed Specific Examples Using a Wikipedia derived resource to (manually or automatically) create training data Compiling lists of name variations based on hyperlinks and redirects prior to the evaluation Using a Wikipedia derived resource before evaluation to create a KB of world knowledge that can be used to check the correctness of facts Preprocess/annotate a large text corpus before the evaluation to check the correctness of facts or aliases Editing Wikipedia pages for target entities, either during, or after the evaluation Table 5. Rules for Using External Resources 5

4.2 Resource Sharing To support groups that intend to focus on part of the tasks, participants are encouraged to share external resources that they prepared before the evaluation. Such resources may include intermediate results, entity annotations, parsing/srl/ie annotated Wikipedia corpus, topic model features for entity linking, etc. 5 Submissions 5.1 Submissions In KBP 2013 participants will have one week after downloading the data to return their results for each task. Up to five alternative system runs may be submitted by each team for each task. Submitted runs should be ranked according to their expected score (based on development data, for example). Systems should not be modified once queries are downloaded. Details about submission procedures will be communicated to the track mailing list. The tools to validate formats will be made available at: http://www.nist.gov/tac/2013/kbp/entitylinking/tools.html. 5.2 Schedule Please visit the KBP 2013 entity-linking website for the schedule for the entity-linking tasks at KBP 2013: http://www.nist.gov/tac/2013/kbp/entitylinking/ 6 Mailing List and Website The KBP 2013 website is http://www.nist.gov/tac/2013/kbp/. The website dedicated to the Entity-Linking tasks is http://www.nist.gov/tac/2013/kbp/entitylinking/. Please post any questions and comments to the mailing list tac-kbp@nist.gov. Information about subscribing to the list is available at: http://www.nist.gov/tac/2013/kbp/registration.html. Change History Version 1.0 o Original version, based on the 2012 specification o Added confidence values and top N scoring 6