- From: Giuseppe Rizzo <giuseppe.rizzo@eurecom.fr>
- Date: Wed, 17 Dec 2014 13:32:59 +0100
- To: public-microposts@w3.org
====================================================================== Named Entity rEcognition and Linking (NEEL) Challenge at the 5th Making Sense of Microposts Workshop #Microposts2015 @ WWW 2015 http://www.scc.lancs.ac.uk/microposts2015/challenge/index.html 18th/19th May 2015, Florence, Italy ====================================================================== Microposts are a highly popular medium to share facts, opinions or emotions. They are an invaluable wealth of data, ready to be mined for training predictive modellings. Following the success of the last year, we are pleased to announce the NEEL challenge which will be part of the #Microposts2015 Workshop at the World Wide Web 2015 conference. The overall task of the challenge is to automatically recognise entities and their types from English microposts, and to link them to the corresponding English DBpedia 2014 resources (if the linkage exists). Participants will have to automatically extract expressions that are formed by discrete (and typically short) sequences of words (e.g., Obama, London, Rakuten) and recognise their types (e.g., Person, Location, Organisation) from a collection of microposts. As linking stage we aim to disambiguate the spotted entity to the corresponding DBpedia resource, or to a NIL reference if the spotted named entity does not match any resource in DBpedia. This year challenge will also evaluate the end-to-end performance of the system by measuring the computation time for analyzing the corpus using the submitted algorithms. We welcome and hope participants from NEEL, TREC, TAC KBP, ERD shared tasks to participate in this year challenge. DATASET ------- The dataset comprises tweets extracted from a collection of over 18 million tweets. The dataset includes event-annotated tweets provided by the Redites project (http://demeter.inf.ed.ac.uk/redites/) covering multiple noteworthy events from 2011, 2013 (including the death of Amy Winhehouse, the London Riots, the Oslo bombing and the Westgate Shopping Mall shootout) and tweets extracted from the Twitter firehose from 2014. Since the task of this challenge is to automatically recognise and link entities, we have built our dataset considering both event and non-event tweets. While event tweets are more likely to contain entities, non-event tweets enable us to evaluate the performance of the system in avoiding false positives in the entity extraction phase. The training set is built on top of the entire corpus of the NEEL 2014 Challenge. We have further extended it for typing the entities and adding the NIL references. Following the Twitter TOS we will only provide tweet IDs and annotations for the training set; and tweet IDs for the test set. We will also provide a common framework to mine these datasets from Twitter. The training set will be released as tsv following the TAC KBP format, where each line consists of the following features: 1st: tweet id 2nd,3rd: start/end offsets expressed as the number of UTF8 characters starting from 0 (the beginning of the tweet) 4th: link to DBpedia resource or NIL (it may exist different NIL in the corpus. Each NIL may be reused if there are multiple mentions in the text which represent the same entity) 5th: salience (confidence score) 6th: type Tokens are separated by TABs. Entity mentions and URIs are listed according to their appearance order in the tweet. We will timely advertise the release of the data sets on the workshop mailing list. Please subscribe to https://groups.google.com/d/forum/microposts2015. EVALUATION ---------- Participants are required to implement their systems as a publicly accessible web service following a REST based protocol (which will be advertised on the mailing list before the release of the training set) and submit their contending entries (up to 10) to a registry of the NEEL challenge services. Upon receiving the registration of the service, calls to the contending entry will be scheduled in two different time windows, namely D-Time (meant to test the APIs) and T-Time for the final evaluation and metric computations. In the final stage, each participant can submit up to 3 final contending entries. We will use the metrics proposed by TAC KBP 2014 (https://github.com/wikilinks/neleval/wiki/Evaluation) and in particular we will focus on: [tagging] strong_typed_mention_match (check entity name boundary and type) [linking] strong_mention_match [clustering] mention_ceaf (NIL detection) To ensure the correctness of the results and avoid any loss we will trigger N number of calls and we will statistically evaluate the metrics. PAPER SUBMISSION ---------------- A paper of 2 pages describing your approach, how you tuned/tested it using the training split, and your results. All submissions must be in English. All written submissions should be prepared according to the ACM SIG Proceedings Template (see http://www.acm.org/sigs/publications/proceedings-templates), and should include author names and affiliations, and 3-5 author-selected keywords. Where a submission includes additional material submission this should be made as a single, unencrypted zip file that includes a plain text file listing its contents.Submission is via EasyChair, at: https://www.easychair.org/conferences/?conf=microposts2015. Each submission will receive at least 2 peer reviews. The #Microposts2015 proceedings will be published as a single volume containing all three tracks, via CEUR. The same publication conditions however apply as for other workshop proceedings included in the WWW conference companion: "Any paper published by the ACM, IEEE, etc. which can be properly cited constitutes research which must be considered in judging the novelty of a WWW submission, whether the published paper was in a conference, journal, or workshop. Therefore, any paper previously published as part of a WWW workshop must be referenced and suitably extended with new content to qualify as a new submission to the Research Track at the WWW conference." WORKSHOP STRUCTURE ------------------ A keynote address from an invited speaker will open the day, and followed by paper presentations. We will hold a poster and demo session to trigger further, in-depth interaction between workshop participants. The last set of presentations will be brief overviews of selected submissions to the Challenge. The workshop will close with the presentation of awards. IMPORTANT DATES --------------- Intent to participate: 20 Jan 2015 (soft - further instructions will be shared on the mailing list ) Release of the REST API specs: 2 Feb 2015 Release of training set: 15 Feb 2015 Registration of contending entries: 2 Mar 2015 D-Time: 10-15 Mar 2015 (hard) T-Time: 20-25 Mar 2015 (hard) Paper submission: 28 Mar 2015 (hard) Challenge Notification: 21 Apr 2015 (hard) Challenge camera-ready deadline: 31 Apr 2015 (hard) Workshop program issued: 22 Apr 2015 Challenge proceedings to be published via CEUR Workshop - 18/19 May 2015 (Registration open to all) (All deadlines 23:59 Hawaii Time) PRIZE ----- A prize of 1500 euros, generously sponsored by SpazioDati, will be awarded to the challenge winner. SpazioDati is an Italian startup focused on text analytics and big data. One the SpazioDati's key components is DataTXT, a text-analytics engine available on SpazioDati API platform, Dandelion. DataTXT named-entity extraction system has been proven to be very effective and efficient on short and fragmented texts, like microposts. By teaming up with SpazioDati to make the challenge possible, the #Microposts workshop organisers wish to highlight new entity extraction methods and algorithms to pursue in such a challenging scenario. CONTACT ------- Mailing list : https://groups.google.com/forum/#!forum/microposts2015 E-mail: microposts2015@easychair.org Twitter hashtags: #neel #microposts2015 Twitter account: @Microposts2015 W3C Microposts Community Group: http://www.w3.org/community/microposts Challenge Organizers: --------------------- Challenge Chair: A. Elizabeth Cano, Knowledge Media Institute, The Open University, UK Giuseppe Rizzo, EURECOM, France Dataset Chair: Andrea Varga, Swiss Re, UK Challenge Committee: --------------------- Gabriele Antonelli, SpazioDati, Italy Ebrahim Bagheri, Ryerson University, Canada Pierpaolo Basile, University of Bari, Italy Leon Derczynski, The University of Sheffield, UK Milan Dojchinovski, Czech Technical University, Czech Republic Guillaume Ereteo, Vigiglobe, France Andrés García-Silva, Univesidad Politécnica de Madrid, Spain Anna Lisa Gentile, Sheffield, UK Miguel Martinez-Alvarez, Signal, UK Jose M. Morales-Del-Castillo, University of Granada, Spain Georgios Paltoglou, University of Wolverhampton, UK Bernardo Pereira Nunes, PUC-Rio, Brazil Daniel Preoţiuc-Pietro, University of Pennsylvania, USA Giles Reger, The University of Manchester, UK Irina Temnikova, Qatar Computing Research Institute, Qatar Raphaël Troncy, EURECOM, France Victoria Uren, Aston University, UK -- Giuseppe Rizzo EURECOM, Multimedia Communications Department 450 Route des Chappes, 06410 Biot, France e-mail: giuseppe.rizzo@eurecom.fr Tel: +33 (0)4 - 9300 8148 Fax: +33 (0)4 - 9000 8200 Web: http://www.eurecom.fr/~rizzo
Received on Wednesday, 17 December 2014 12:34:58 UTC