- From: Filip Ilievski <filip.dbrsk@gmail.com>
- Date: Wed, 7 Jun 2017 12:51:21 -0400
- To: semantic-web@w3.org
- Message-ID: <CANbunvijrk=PpXMqw_JRqVxhsoQF7sR53TPn2A9tFr4mykFr7g@mail.gmail.com>
It is a pleasure to invite you to the workshop SLT-1: Semantics of the Long Tail <https://www.google.com/url?q=http://www.understandinglanguagebymachines.org/semantics-of-the-long-tail/&sa=D&ust=1494255281396000&usg=AFQjCNFochPRp7OrC3VbVM4RzzET9iVxNg>. The workshop is co-located with the IJCNLP 2017 <https://www.google.com/url?q=http://ijcnlp2017.org&sa=D&ust=1494255281397000&usg=AFQjCNFasdrWAuDFo747FTmoHAIFwd47tg> conference, which is organized by The Asian Federation of Natural Language Processing. IJCNLP 2017 will take place in Taipei, Taiwan from the 27th of November to the 1st of December 2017. The use of language expressions and their meanings follow a Zipfian distribution, featuring a small amount of very frequent observations and a very long tail of low frequent observations. As a result of this, the data we use for learning and testing systems also exhibit overfitting to the head of the distribution. It is thus not surprising that statistical approaches automatically exploit the distributional preference and dominance of the most ‘popular’ interpretations for disambiguation and reference tasks in NLP: most of the test interpretations in a task correspond to the majority interpretations of the training instances. But what about the long-tail cases? How well are systems capable of semantically interpreting less and low-frequent cases? Surprisingly, humans do not experience significant problems determining that a long-tail interpretation applies in a specific text. But how can semantic NLP systems intelligently deal with sparse cases? Little attention has been devoted to how systems should solve interpretation tasks for local and perhaps unknown event and entity instances, which are described in sparse documents, but might not be observed in any training set or knowledge base. Potentially, this would require new representations and deeper processing than the ones that work well on the head, which involves reading between the lines, e.g. textual entailment and robust (common sense) reasoning. It is the goal of this workshop to bring researchers together to share their experiences and results in dealing with long tail semantics: i.e. interpretation of low-frequent or rare events, entities and relations for which there is little training data. We welcome papers that describe: (error) analyses of semantic tasks with respect to head and long-tail distributions, evaluation of different methods on head and long-tail cases, new methods to interpret long-tail phenomena, and the role of context to prime long-tail cases over head interpretations. We are interested in how knowledge and data can be acquired to counterbalance popular data and interpretations and how to make semantic tasks sustainable over time when the world and the interpretation space changes. Topics of interest Topics include, but are not limited to, the following: 1. System Performance 1. How can we define the head and the tail for each semantic NLP task? 2. Which evaluation metrics are needed to gain insight into system performance on the tail? 3. Do existing datasets suffice to gain insight into tail performance? What kind of benchmarks are needed to better track progress in processing the tail? 1. Data and Knowledge Requirements 1. What kind and amount of (annotated) data is needed? 2. Do customary knowledge sources (e.g. DBpedia, BabelNet, and WordNet) suffice? 3. Do we need massive local knowledge resources to represent the world and all its contexts? 1. Methods and linguistic representations 1. Are the methods and representations needed for the tail different from the ones for the head? 2. How can we transfer models developed for the head to make them appropriate for modeling the tail? 3. How can the recent advances in deep neural networks and matrix factorization be directed to accomplish this? 1. Contextual adaptation 1. How to build systems that can switch between contexts of time, topic, and location (e.g. how to build systems that can adapt to new or past long tail realities)? Guide for authors The deadline to submit papers is September 5, 2017. Paper submissions for IJCNLP will be handled by the Softconf START system (submission link will be provided once available). The program chairs will release both Latex and Microsoft Word templates soon. Prospective authors should submit an extended abstract of 2 pages in length, excluding references, and will be asked to extend it to a long or short paper upon acceptance. For more information on short papers (5 pages of content + 2 pages for references) and long papers (9 pages of content + 2 pages for references), we refer to: http://ijcnlp2017.org/site /page.aspx?pid=148&sid=1133&lang=en <https://www.google.com/url?q=http://ijcnlp2017.org/site/page.aspx?pid%3D148%26sid%3D1133%26lang%3Den&sa=D&ust=1494255281412000&usg=AFQjCNGDvsG-y2sdBbPf6DrNqTgUAOy3qQ> Important Dates Deadline for submission: September 5, 2017 Notification of acceptance: September 30, 2017 Deadline for final paper submission: October 10, 2017 Workshop Chairs Initiators: Piek Vossen (Vrije Universiteit Amsterdam) Marten Postma (Vrije Universiteit Amsterdam) Filip Ilievski (Vrije Universiteit Amsterdam) Co-organizers: Martha Palmer (University of Colorado Boulder) Chris Welty (Google) Ivan Titov (University of Edinburgh) Eduard Hovy (Carnegie Mellon University) Eneko Agirre (University of the Basque Country) Philipp Cimiano (University of Bielefeld) Frank van Harmelen (Vrije Universiteit Amsterdam) Key-Sun Choi (Korea Advanced Institute of Science and Technology) For any enquiries regarding the programme, please contact: m.c.postma@vu.nl or f.ilievski@vu.nl For all general enquiries, please contact: m.c.postma@vu.nl or f ..ilievski@vu.nl We look forward to seeing you at SLT-1. Sincerely, The Organizing committee of SLT-1
Received on Friday, 9 June 2017 09:28:36 UTC