TAC KBP Chinese Cross-lingual Entity Linking Comprehensive Training and Evaluation Data 2011-2014 Authors: Joe Ellis, Jeremy Getman, Stephanie Strassel 1. Overview This package contains training and evaluation data produced in support of the TAC KBP Chinese Cross-lingual Entity Linking evaluation tracks from 2011 to 2014. Text Analysis Conference (TAC) is a series of workshops organized by the National Institute of Standards and Technology (NIST). TAC was developed to encourage research in natural language processing (NLP) and related applications by providing a large test collection, common evaluation procedures, and a forum for researchers to share their results. Through its various evaluations, the Knowledge Base Population (KBP) track of TAC encourages the development of systems that can match entities mentioned in natural texts with those appearing in a knowledge base and extract novel information about entities from a document collection and add it to a new or existing knowledge base. Chinese cross-lingual Entity Linking was first conducted as part of the 2011 TAC KBP evaluations. The track was an extension of the monolingual English Entity Linking track (EL) whose goal is to measure systems' ability to determine whether an entity, specified by a query, has a matching node in a reference knowledge base (KB) and, if so, to create a link between the two. If there is no matching node for a query entity in the KB, EL systems are required to cluster the mention together with others referencing the same entity. More information about the TAC KBP Entity Linking task and other TAC KBP evaluations can be found on the NIST TAC website, http://www.nist.gov/tac/. This package contains all evaluation and training data developed in support of TAC KBP Chinese Cross-lingual Entity Linking during the four years in which the task was conducted, from 2011-2014. This includes queries and gold standard entity type information, KB links, and equivalence class clusters for NIL entities (those for which there was no matching node in the KB). Source documents for the queries are included in this corpus. The corresponding KB is available as LDC2014T16: TAC KBP Reference Knowledge Base. The data included in this package were originally released to TAC KBP as: LDC2011E46: TAC 2011 KBP Cross-lingual Sample Entity Linking Queries V1.1 LDC2011E55: TAC 2011 KBP Cross-lingual Training Entity Linking V1.1 LDC2012E34: TAC 2011 KBP Cross-Lingual Evaluation Entity Linking Annotation LDC2012E66: TAC 2012 KBP Chinese Entity Linking Web Training Queries and Annotations LDC2012E103: TAC 2012 KBP Chinese Entity Linking Evaluation Annotations V1.2 LDC2013E96: TAC 2013 KBP Chinese Entity Linking Evaluation Queries and Knowledge Base Links V1.2 LDC2014E47: TAC 2014 KBP Chinese Entity Linking Discussion Forum Training Data LDC2014E83: TAC 2014 KBP Chinese Entity Linking Evaluation Queries and Knowledge Base Links V2.0 LDC2015E17: TAC KBP Chinese Entity Linking Comprehensive Training and Evaluation Data 2011-2014 Summary of data included in this package (for more details see /docs/tac_kbp_2011-2014_chinese_entity_linking_query_distribution_table.tsv): +------+------------------+---------+ | Year | Source Documents | Queries | +------+------------------+---------+ | 2011 | 4329 | 4347 | | 2012 | 2271 | 2280 | | 2013 | 2143 | 2155 | | 2014 | 2860 | 3253 | +------+------------------+---------+ 2. Contents ./README.txt This file ./data/{2011,2012,2013,2014}/contents.txt The data in this package are organized by the year of original release in order to clarify dependencies, highlight occassional differences in formats from one year to another, and to increase readability in documentation. The contents.txt file within each year's root directory provides a list of the contents for all subdirectories as well as details about file formats and contents. ./docs/all_files.md5 Paths (relative to the root of the corpus) and md5 checksums for all files in the package. ./docs/tac_kbp_2011-2014_chinese_entity_linking_query_distribution_table.tsv Tab-delimited table containing the query distribution quantities for all years and datasets, further broken down by language, source type, KB-Link, and entity type. ./docs/guidelines/2011/* The guidelines used by annotators in developing the 2011 Entity Linking queries and gold standard data contained in this corpus. ./docs/guidelines/2012/TAC_KBP_2012_Entity_Selection_V1.1.pdf The guidelines used by annotators in developing the 2012 Entity Linking queries and gold standard data contained in this corpus. ./docs/guidelines/2013/TAC_KBP_2013_EL_Query_Development_Guidelines_V1.0.pdf The guidelines used by annotators in developing the 2013 Entity Linking queries and gold standard data contained in this corpus. ./docs/guidelines/2014/TAC_KBP_2014_EL_Query_Development_Guidelines_V1.0.pdf The guidelines used by annotators in developing the 2014 Entity Linking queries and gold standard data contained in this corpus. ./docs/task_descriptions/KBP2011_TaskDefinition.pdf Task Description for all of the 2011 TAC KBP tracks, written by track coordinators. ./docs/task_descriptions/KBP2012_TaskDefinition_1.1.pdf Task Description for all of the 2012 TAC KBP tracks, written by track coordinators. ./docs/task_descriptions/KBP2013_EntityLinkingTaskDescription_1.0.pdf Task Description for the 2013 Entity Linking evaluation tracks, written by track coordinators. ./docs/task_descriptions/KBP2014EL_V1.1.pdf Task Description for the 2014 Entity Linking evaluation tracks, written by track coordinators. ./dtd/clel_queries_2011.dtd DTD for: ./data/2011/eval/tac_kbp_2011_chinese_entity_linking_evaluation_queries.xml ./data/2011/training/tac_kbp_2011_chinese_entity_linking_sample_and_training_queries.xml ./dtd/clel_queries_2012-2014.dtd DTD for: ./data/2012/eval/tac_kbp_2012_chinese_entity_linking_evaluation_queries.xml ./data/2012/training/tac_kbp_2012_chinese_entity_linking_training_queries.xml ./data/2013/eval/tac_kbp_2013_chinese_entity_linking_evaluation_queries.xml ./data/2014/eval/tac_2014_kbp_chinese_entity_linking_evaluation_queries.xml ./data/2014/training/tac_kbp_2014_chinese_entity_linking_training_queries.xml ./tools/check_kbp_2011_cross-lingual-entity-linking.pl Validator for 2011 entity linking submission files, as provided to LDC by evaluation track coordinators, with no further testing. ./tools/check_kbp2012_chinese-entity-linking.pl Validator for 2012 entity linking submission files, as provided to LDC by evaluation track coordinators, with no further testing. ./tools/check_kbp2013_2014_chinese-entity-linking.pl Validator for 2013 and 2014 entity linking submission files, as provided to LDC by evaluation track coordinators, with no further testing. ./tools/el_scorer_2011_2012.py Scorer for 2011 and 2012 entity linking submission files, as provided to LDC by evaluation track coordinators, with no further testing. ./tools/el_scorer_2013.py Scorer for 2013 entity linking submission files, as provided to LDC by evaluation track coordinators, with no further testing. ./tools/el_scorer_2014.py Scorer for 2014 entity linking submission files, as provided to LDC by evaluation track coordinators, with no further testing. 3. Query Development Annotation and Quality Control Query development for Entity Linking begins with Entity Selection, which has three stages: Namestring Annotation, Knowledge Base (KB) Linking, and NIL Coreference (where a NIL entity is an entity without a node in the KB). Bilingual Chinese/English-speaking annotators searched the corpus for entities that would make suitable queries, using an interface created by LDC for this task. Each set of queries was roughly balanced across entity type, KB-link status(NIL vs. non-NIL), and source document genre. Most queries were drawn from non-English documents, but mentions in English documents of entities co-referential with other non-English queries were selected whenever possible. In Namestring Annotation, annotators search for and select named mentions of entities in text. Annotators focused on creating queries using confusable named entity mentions. Confusability was measured both by the number of distinct entities in the full query set referred to by the same name string (polysemy) as well as the number of distinct entities in the set that were referred to by multiple, unique named mentions (synonymy). For example, the string "Smith" would make a polysemous query because an annotator could probably find it in the corpus referring to different entities, while "Barack Obama" would make a synonymous query because the entity is also referred to in the corpus as "B. Hussein Obama" or "Bam Bam". In KB Linking, annotators search the KB and indicate whether or not it includes pages on the entities they selected during Namestring Annotation. Annotators created a link between the query and the matching KB node ID. If no matching node was found, the query was marked as NIL and later coreferenced with other NIL entities. Annotators were allowed to use online searching to assist in determining the KB link/NIL status. Queries for which an annotator could not confidently determine the KB link status were removed from the final data sets. For NIL Coreference, selected entities that were not included in the KB (i.e., NIL entities) were grouped into equivalence classes by annotators. Mentions referring to the same entity were grouped into one equivalence class. Senior annotators conducted quality control on queries to correct errors and identify areas of difficulty for use in improving guidelines and annotator training. Annotators performing quality control made sure that the extent of each selected namestring was correct and checked that each entity was linked to the correct KB node or was properly identified as NIL and coreferenced correctly. 4. Source Documents The source data contained in this release comprises all documents from which queries were drawn and is the complete data set used in the Chinese EL evaluations. The source documents were drawn from existing LDC holdings, with no additional validation. An overall scan of character content in the source collections indicates some relatively small quantities of various problems, especially in the web and discussion forum data, including language mismatch (characters from Chinese, Korean, Japanese, Arabic, Russian, etc.), and encoding errors (some documents have apparently undergone "double encoding" into UTF-8, and others may have been "noisy" to begin with, or may have gone through an improper encoding conversion, yielding occurrences of the Unicode "replacement character" (U+FFFD) throughout the corpus); the web collection also has characters whose Unicode code points lie outside the "Basic Multilanguage Plane" (BMP), i.e. above U+FFFF. All source documents were originally released as XML but have been converted to text files for this release. This change was made primarily because the documents were used as text files during data development but also because some fail XML parsing. All documents that have filenames beginning with "eng-NG" are Web Document data (WB) and some of these fail XML parsing (see below for details). All files that start with "bolt-" are Discussion Forum threads (DF) and have the XML structure described below. All other files are Newswire data (NW) and have the newswire markup pattern detailed below. Note as well that some source documents are duplicated across a few of the separated source_documents directories, indicating that some queries from different data sets originated from the same source documents. As it is acceptable for source to be reused for Entity Linking queries, this duplication is intentional and expected. The subsections below go into more detail regarding the markup and other properties of the three source data types: 4.1 Newswire Data Newswire data use the following markup framework: ... ...

...

...
where the HEADLINE and DATELINE tags are optional (not always present), and the TEXT content may or may not include "

...

" tags (depending on whether or not the "doc_type_label" is "story"). All the newswire files, if converted back to XML, are parseable. 4.2 Discussion Forum Data Discussion forum files use the following markup framework: ... ... ... ... ... where there may be arbitrarily deep nesting of quote elements, and other elements may be present (e.g. "..." anchor tags). As mentioned in section 2 above, each unit contains at least five post elements. All the discussion forum files, if converted back to XML, are parseable. 4.3 Web Document Data "Web" files use the following markup framework: {doc_id_string} ... ... ... ... ... ... Other kinds of tags may be present ("", "", etc). Some of the web source documents contain material that interferes with XML parsing (e.g. unescaped "&", or "" tags that lack a corresponding ""). 5. Using the Data 5.1 Offset calculation The values of the and XML elements in the later queries.xml files indicate character offsets to identify text extents in the source. Offset counting starts from the initial opening angle bracket of the element ( in DF sources), which is usually the initial character (character 0) of the source. Note as well that character counting includes newlines and all markup characters - that is, the offsets are based on treating the source document file as "raw text", with all its markup included. 5.2 Proper ingesting of XML queries While the character offsets are calculated based on treating the source document as "raw text", the "name" strings being referenced by the queries sometimes contain XML metacharacters, and these had to be "re-escaped" for proper inclusion in the queries.xml file. For example, an actual name like "AT&T" may show up a source document file as "AT&T" (because the source document was originally formatted as XML data). But since the source doc is being treated here as raw text, this name string is treated in queries.xml as having 7 characters (i.e., the character offsets, when provided, will point to a string of length 7). However, the "name" element itself, as presented in the queries.xml file, will be even longer - "AT&T" - because the queries.xml file is intended to be handled by an XML parser, which will return "AT&T" when this "name" element is extracted. Using the queries.xml data without XML parsing would yield a mismatch between the "name" value and the corresponding string in the source data. 6. Acknowledgements This material is based on research sponsored by Air Force Research Laboratory and Defense Advance Research Projects Agency under agreement number FA8750-13-2-0045. The U.S. Government is authoized to reporoduce and distribute reprints for Governmental purposes notwithstanding any copyright notation thereon. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of Air Force Research Laboratory and Defense Advanced Research Projects Agency or the U.S. Government. The authors acknowledge the following contributors to this data set: Dave Graff (LDC) Neil Kuster (LDC) Xuansong Li (LDC) Kira Griffitt (LDC) Heng Ji (RPI) Hoa Dang (NIST) Ralph Grishman (NYU) Javier Artiles (Slice Technologies) Boyan Onyshkevych (DARPA) 7. References Joe Ellis, Jeremy Getman, Stephanie M. Strassel. 2014 Overview of Linguistic Resources for the TAC KBP 2014 Evaluations: Planning, Execution, and Results https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/tackbp-2014-overview.pdf TAC KBP 2014 Workshop: National Institute of Standards and Technology, Gaithersburg, Maryland, November 17-18 Joe Ellis, Jeremy Getman, Justin Mott, Xuansong Li, Kira Griffitt, Stephanie M. Strassel, Jonathan Wright. 2013 Linguistic Resources for 2013 Knowledge Base Population Evaluations https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/tackbp-workshop2013-linguistic-resources-kbp-eval.pdf TAC KBP 2013 Workshop: National Institute of Standards and Technology, Gaithersburg, MD, November 18-19 Joe Ellis, Xuansong Li, Kira Griffitt, Stephanie M. Strassel, Jonathan Wright. 2012 Linguistic Resources for 2012 Knowledge Base Population Evaluations https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/tackbp-workshop2012-linguistic-resources-kbp-eval.pdf TAC KBP 2012 Workshop: National Institute of Standards and Technology, Gaithersburg, MD, November 5-6 Xuansong Li, Joe Ellis, Kira Griffit, Stephanie Strassel, Robert Parker, Jonathan Wright. 2011 Linguistic Resources for 2011 Knowledge Base Population Evaluation https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/tac2011-linguistic-resources-kbp.pdf TAC 2011: Proceedings of the Fourth Text Analysis Conference, Gaithersburg, Maryland, November 14-15 8. Copyright Information (c) 2016 Trustees of the University of Pennsylvania 9. Contact Information For further information about this data release, contact the following project staff at LDC: Joseph Ellis, Project Manager Jeremy Getman, Lead Annotator Stephanie Strassel, PI -------------------------------------------------------------------------- README created by Neil Kuster on November 17, 2015 updated by Joe Ellis on November 20, 2015 updated by Dana Fore on February 3, 2016 updated by Joe Ellis on April 22, 2016