A Framework for Benchmarking Entity-Annotation Systems
11 pages
English

Découvre YouScribe en t'inscrivant gratuitement

Je m'inscris

A Framework for Benchmarking Entity-Annotation Systems

Découvre YouScribe en t'inscrivant gratuitement

Je m'inscris
Obtenez un accès à la bibliothèque pour le consulter en ligne
En savoir plus
11 pages
English
Obtenez un accès à la bibliothèque pour le consulter en ligne
En savoir plus

Description

AFrameworkforBenchmarkingEntity-Annotation Systems Marco Cornolti Paolo Ferragina Massimiliano Ciaramita Dipartimento di Informatica Dipartimento di Informatica Google Research University of Pisa, Italy University of Pisa, Italy Zürich, Switzerland massi@google.comcornolti@di.unipi.it ferragina@di.unipi.it ABSTRACT ture but low coverage (such as WordNet, CYC, TAP), and a large text collection with wide coverage but unstructuredIn this paper we design and implement a benchmarking and noisy content (like the whole Web). The process of en-framework for fair and exhaustive comparison of entity-an- tity annotation involves three main steps: (1) parsing of thenotation systems. The framework is based upon the de - input text, which is the task to detect candidate entity men-nition of a set of problems related to the entity-annotation tions and link each of them to all possible entities they couldtask, a set of measures to evaluate systems performance, mention; (2) disambiguation of mentions, which is the taskand a systematic comparative evaluation involving all pub- of selecting the most pertinent Wikipedia page (i.e., entity)licly available datasets, containing texts of various types that best describes each mention; (3) pruning of a mention,such as news, tweets and Web pages.

Sujets

Informations

Publié par
Publié le 19 mai 2013
Nombre de lectures 39
Langue English

Extrait

AFrameworkforBenchmarkingEntity-Annotation
Systems
Marco Cornolti Paolo Ferragina Massimiliano Ciaramita
Dipartimento di Informatica Dipartimento di Informatica Google Research
University of Pisa, Italy University of Pisa, Italy Zürich, Switzerland
massi@google.comcornolti@di.unipi.it ferragina@di.unipi.it
ABSTRACT ture but low coverage (such as WordNet, CYC, TAP), and a
large text collection with wide coverage but unstructuredIn this paper we design and implement a benchmarking
and noisy content (like the whole Web). The process of en-framework for fair and exhaustive comparison of entity-an-
tity annotation involves three main steps: (1) parsing of thenotation systems. The framework is based upon the de-
input text, which is the task to detect candidate entity men-nition of a set of problems related to the entity-annotation
tions and link each of them to all possible entities they couldtask, a set of measures to evaluate systems performance,
mention; (2) disambiguation of mentions, which is the taskand a systematic comparative evaluation involving all pub-
of selecting the most pertinent Wikipedia page (i.e., entity)licly available datasets, containing texts of various types
that best describes each mention; (3) pruning of a mention,such as news, tweets and Web pages. Our framework is
which discards a detected mention and its annotated entityeasily-extensible with novel entity annotators, datasets and
if they are considered not interesting or pertinent to theevaluation measures for comparing systems, and it has been
1 semantic interpretation of the input text.released to the public as open source . We use this frame-
The focus around entity annotators has increased signif-work to perform the rst extensive comparison among all
icantly in the last few years, with several interesting andavailable entity annotators over all available datasets, and
e ective algorithmic approaches to solve the mention-entitydraw many interesting conclusions upon their e ciency and
match problem, possibly using other knowledge bases suche ectiveness. We also draw conclusions between academic
as DBpedia, Freebase or Yago (see e.g. [2, 3, 7, 9, 13, 16,versus commercial annotators.
18, 19, 23, 5, 15]). Unfortunately, the research has inves-
tigated some speci c tasks using non-uniform terminology,
Categories and Subject Descriptors non-comparable evaluation metrics, and limited datasets and
D.2.8 [Software Engineering]: Metrics|performance mea- systems. As a consequence, we only have a partial picture of
the e ciency and e ectiveness of known annotators whichsures; I.2.7 [Arti cial Intelligence ]: Natural Language
Processing|text analysis makes it di cult to compare them in a fair and complete
way. This is a particularly important issue because those
systems are being used as black-boxes by more and more IRKeywords
tools, built on top of them, such as [1, 8, 21, 22]. Motivated
Benchmark Framework; Entity annotation; Wikipedia by these considerations, [20] recently attempted to compare
entity-annotation systems mainly coming from the commer-
21. INTRODUCTION cial realm . However, as the authors state in the concluding
section of their paper, their evaluation is limited to strictClassic approaches to document indexing, clustering, clas-
metrics which account for \exact" matches over mentionssication and retrieval are based on the bag-of-words para-
and entities (and, rather observe that \a NE type might bedigm. The limitations of this paradigm are well-known to
not wrong but not precise enough."), it does not considerthe IR community and in recent years a good deal of work
datasets fully annotated by humans (i.e. mentions are onlyhas attempted to move beyond by\grounding"the processed
the ones derived by few parsers), and misses to consider thetexts with respect to an adequate semantic representation,
best performing tools which have been published recently inby designing so-called entity annotators. The key idea is to
the scienti c literature. This last issue is a crucial limita-identify, in the input text, short-and-meaningful sequences
tion because, as [12] showed recently, the DBpedia Spotlightof terms (also called mentions) and annotate them with un-
system (the best according to [20]) achieves much worse per-ambiguous identiers (also called entities) drawn from a cat-
formance than some of the systems tested in this paper.alog. Most recent work adopts anchor texts occurring in
Given this scenario, we aim with this paper at de ning andWikipedia as entity mentions and the respective Wikipedia
implementing a framework for comparing in a complete, fairpages as the mentioned entity, because Wikipedia oers to-
and meaningful way the most e cient, e ective and pub-day the best trade-o between catalogs with a rigorous struc-
licly available entity-annotation systems: namely, AIDA [7],
1See http://acube.di.unipi.it/ Illinois Wikier [19], TagMe [3], Wikipedia-miner [16], and
CopyrightisheldbytheInternationalWorldWideWebConference 2AlchemyAPI, DBpedia Spotlight, Evri, Extractiv, Lupe-Committee(IW3C2). Distributionofthesepapersislimitedto
dia, OpenCalais, saplo, Wikimeta, Yahoo! Content Analysisclassroomuse,andpersonalusebyothers.
and Zemanta.WWW 2013,May13–17,2013,RiodeJaneiro,Brazil.
ACM978-1-4503-2035-1/13/05.Problem Input Output Description
Disambiguate to Wikipedia (D2W) Text, Set of relevant annota- Assign to each input mention its pertinent entity (possibly
Set of tionsw null). This problem has been introduced in [2].
mentions
Annotate to Wikipedia (A2W) Text Set of relevant annota- Identify the relevant mentions in the input text and assign
tions to each of them the pertinent entities. This problem has
been introduced in [16].
Scored-annotate to Wikipedia (Sa2W) Text Set of relevant and scored As A2W, but here each annotation is assigned a score
annotations representing the likelihood that the annotation is correct.
This problem has been introduced in [16].
Concepts to Wikipedia (C2W) Text Set of relevant tags Tags are taken as the set of relevant entities that are men-
tioned in the input text. This problem has been dened
in [13].
Scored concepts to Wikipedia (Sc2W) Text Set of relevant and scored As C2W, but here each tag is assigned a score representing
tags. the likelihood that the annotation is correct.
Ranked-concepts to Wikipedia (Rc2W) Text Ranked list of relevant Identify the entities mentioned in a text and rank them
tags in terms of their relevance for the topics dealt with in the
input text. This problem has been dened in [13].
Table 1: A set of entity-annotation problems.
Problem Input OutputDBpedia Spotlight; which currently dene the state-of-the-
art for the entity-annotation task. In order to achieve this The novel begins in the Shire, where the The novel begins in the Shire, where theD2W
Hobbit Frodo Baggins inherits the Ring Hobbit Frodo Baggins inherits the Ring
goal, we will introduce (1) a hierarchy of entity-annotation from Bilbo from Bilbo
NULLproblems, that cover the wide spectrum of annotation goals
Novelsuch systems could address; (2) a set of novel measures to

  • Univers Univers
  • Ebooks Ebooks
  • Livres audio Livres audio
  • Presse Presse
  • Podcasts Podcasts
  • BD BD
  • Documents Documents