[open-linguistics] CfP: Entity Extraction and Linking Challenge (#Microposts2014 @ WWW2014)

ampaeli cano ampaeli at gmail.com
Sat Dec 21 14:49:39 UTC 2013


=======================================================================
              Entity Extraction and Linking Challenge
           at the 4th Making Sense of Microposts Workshop
                   (#Microposts2014) @ WWW 2014
   http://www.scc.lancs.ac.uk/microposts2014/challenge/index.html
            7 April 2014, Seoul, Republic of Korea
=======================================================================

Microposts are a highly popular medium to share facts, opinions or
emotions. They are an invaluable wealth of data, ready to be mined for
training predictive modelings. This year the #Microposts 2014 Workshop will
host an "Entity Extraction and Linking Challenge".
The overall task of the challenge is to automatically extract entities from
English microposts, and link them to the corresponding English DBpedia v3.9
resources (if the linkage exists). As linking stage we aim to to
disambiguate
expressions that are formed by discrete (and typically short) sequences of
words.
Existing entity linking tools are intended for use over news corpora and
similar document-based corpora with relatively long length. We organise
this challenge to foster research into novel, more accurate solutions for
the automatic entity linking in (much shorter) micropost data.
We will ask the participants to automatically extract entities (e.g.,
Obama, London, Rakuten)  belonging to all entity types (e.g., Person,
Location, Organisation) from a collection of microposts. Participants will
have to automatically provide context-relevant DBpedia resources for each
entity in a micropost.

DATASET
-------
The dataset comprises of 3.5K tweets extracted from a much larger
collection of over 18 million tweets. This collection, provided by the
Redites project (http://demeter.inf.ed.ac.uk/redites/), covers
event-annotated tweets collected for the period of 15th July 2011 to 15th
August 2011 (31 days). It extends over multiple noteworthy events including
the death of Amy Winhehouse, the London Riots and the Oslo bombing. Since
the task of this challenge is to automatically extract and link entities,
we have built our dataset considering both event and non-event tweets.
While event tweets are more likely to contain entities, non-event tweets
enable us to evaluate the performance of the system in avoiding false
positives in the entity extraction phase.

The dataset has been split into a training (70%) and testing (30%) sets.
Following the Twitter TOS we will only provide tweet IDs and annotations
for the training set; and tweet IDs for the test set. We will also provide
a common framework to mine these datasets from Twitter.

The training set will be released as tsv file where each line consists of :
- tweet_id
- entity_mention_1
- entity_uri_1
…
- entity_mention_n
- entity_uri_n
Tokens are separated by TABs. Entity mentions and uris are listed according
to their appearance order in the tweet.

We will timely advertise the release of the data sets on the workshop
mailing list. Please subscribe to https://groups.google.com/d/
forum/microposts2014. More information about dates are available in the
Challenge website.

EVALUATION
----------
The evaluation consists of two separated stages:
1.- Paper peer review : A community of experts of the domain will judge the
quality and applicability of the approaches taken, to provide useful
insights on your research;
2.- Precision and Recall:  F1 (F-measure with beta = 1) will be computed on
a gold standard manually created from the test set. The automatically
extracted entities and links will be both matched against this ground truth.

All submissions will be only ranked according to the F1 of each best
submission.

SUBMISSIONS
-----------
Submissions should be provided as a zip file using your system name as the
file name (e.g. 'awesome.zip'), containing:

1. a TSV file with your system name (e.g. 'awesome.tsv'). We accept up to 3
different submissions, and we will consider *only* the best. If you do so
you must specify clearly in your paper the modifications applied to each
labelled submission. In this case the submission should contain each of up
to 3 TSV files with the tool/system name with "_n" appended to each (e.g.
awesome_1.tsv, awesome_2.tsv, awesome_3 ).
In order to evaluate your submissions we require you to submit a tsv file
following the format in which the training set is provided.

2. a paper of 6 pages describing your approach and how you tuned/tested it
using the training split. All submissions must be in English. Submissions
must be in PDF formatted in the style of the Springer Publications format
for Lecture Notes in Computer Science (LNCS) [http://www.springer.com/
computer/lncs?SGWID=0-164-6-793341-0]. For details on the LNCS style, see
Springer’s Author Instructions. All submissions are not anonymous. Please
send us your submission before the deadline through Easychair [
https://www.easychair.org/conferences/?conf=microposts2014]. All accepted
submissions will be invited for short presentations during the
#Microposts2014 workshop and will be published independently from the
workshop proceedings on the challenge page and on CEUR [http://ceur-ws.org/]
(note that a minimum number of papers should be submitted in order to be
able to publish them on CEUR).

IMPORTANT DATES
---------------
Intent to participate: 13 Jan 2014 (soft)
Release of training set: 14 Jan 2014
Release of test set: 17 Feb 2014
Challenge Submission deadline: 21 Feb 2014 (hard)
Challenge Notification: 14 Mar 2014  (hard)
Challenge camera-ready deadline: 24 Mar 2014 (hard)

Workshop program issued: 15 Mar 2014
Challenge proceedings to be published via CEUR
Workshop - 07 Apr 2014 (Registration open to all)
(All deadlines 23:59 Hawaii Time)

PRIZE
-----
to be announced

CONTACT
-------
E-mail: microposts2014 at easychair.org
Facebook Group: http://www.facebook.com/#!/home.php?sk=group_180472611974910
Facebook Public Event page: http://www.facebook.com/events/116134955169543
Google group : https://groups.google.com/forum/#!forum/microposts2014
Twitter hashtag: #microposts2014challenge
Twitter account: @Microposts2014
W3C Microposts Community Group: http://www.w3.org/community/microposts

Challenge Organizers:
------------------------------
Challenge Chair:
A. Elizabeth Cano, Aston University, UK
Giuseppe Rizzo,  Università di Torino, Italy

Dataset  Chair:
Andrea Varga, The University of Sheffield, UK

Challenge  Committee:
---------------------
Ebrahim Bagheri, Ryerson University, Canada
Pierpaolo Basile, Dipartimento di Informatica - University of Bari, Italy
Uldis Bojars, SIOC Project
Óscar Corcho, Universidad Politécnica de Madrid, Spain
Leon Derczynski, The University of Sheffield, UK
Guillaume Erétéo, Orange Labs
Miriam Fernandez, Knowledge Media Institute, The Open University, UK
Andrés García-Silva, Ontology Engineering Group, Facultad de Informática,
Univesidad Politécnica de Madrid, Spain
Anna Lisa Gentile, The University of Sheffield, UK
Robert Jäschke, L3S Research Center, Germany
Diana Maynard,  The University of Sheffield, UK
José M. Morales-Del-Castillo, El Colegio de México, Mexico
Georgios Paltoglou, University of Wolverhampton, UK
Bernardo Pereira Nunes, PUC-Rio, Brazil
Daniel Preoţiuc-Pietro, The University of Sheffield, UK
Raphaël Troncy, EURECOM, France
Mischa Tuffield, PeerIndex
Victoria Uren, Aston University, UK
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.okfn.org/pipermail/open-linguistics/attachments/20131221/609f30f4/attachment-0002.html>


More information about the open-linguistics mailing list