W3C home > Mailing lists > Public > public-vocabs@w3.org > September 2015

[ANNOUNCE] Fact Extraction from Wikipedia Text datasets released

From: Marco Fossati <fossati@fbk.eu>
Date: Wed, 2 Sep 2015 20:29:41 +0200
To: info@iaoa.org, semantic-web@w3.org, public-lod@w3.org, public-ontolex@w3.org, CHI-ANNOUNCEMENTS@listserv.acm.org, aisworld@lists.aisnet.org, planetkr@kr.org, Community@sti2.org, semanticweb@yahoogroups.com, linguist@linguistlist.org, dbpedia-discussion@lists.sourceforge.net, dbpedia-developers@lists.sourceforge.net, public-ldp@w3.org, semantic_web_doktorandennetzwerk@lists.spline.inf.fu-berlin.de, public-vocabs@w3.org, dl@dl.kr.org, spaghettiopendata@googlegroups.com
Message-ID: <55E74015.4010907@fbk.eu>
[Begging pardon if you read this multiple times]

The Italian DBpedia chapter, on behalf of the whole DBpedia Association, 
is thrilled to announce the release of new datasets extracted from 
Wikipedia text.

This is the outcome of an outstanding Google Summer of Code 2015 
project, which implements NLP techniques to acquire structured facts 
from a textual corpus.

The approach has been tested on the soccer use case, with the Italian 
Wikipedia as input.

The datasets are publicly available at:

and loaded into the SPARQL endpoint at:

You can check out this article for more details:

If you feel adventurous, you can fork the codebase at:

Get in touch with Marco at fossati@fbk.eu for everything else.

Best regards,
Marco Fossati
Received on Wednesday, 2 September 2015 18:30:12 UTC

This archive was generated by hypermail 2.4.0 : Friday, 17 January 2020 17:49:41 UTC