- From: Saurav Karmakar <saurav.karmakar@insight-centre.org>
- Date: Thu, 9 Jul 2020 10:24:13 +0100
- To: public-ontolex@w3.org
- Message-ID: <CAFU8_DWW0mt2ZGKVj1dfyKega7MfYJvNRzPgjn9kGcw=_iH_ow@mail.gmail.com>
Dear all, Natural language driven text describes facts, stories, concepts etc and perception of all such happens mostly visually and as well other four senses in reality; visual sense carries a lot in comparison. Adding a visual (image and video) annotation to the ontolex lemon model seems a great value addition, which can definitely provide solidity to natural language grounding and computer vision and its associated fields. There have been a lot of different efforts in natural grounding as well as CVPR but such an standardized is still yet to happen; Imagenet is mentionable in this regard as a is a large visual database mapped with wordnet. Image describes objects or static entities, while videos are essential to describe action or abstract items for its systematic grounding. I am wondering if I can take a stride to focus on this direction as a participant of the public-ontolex group and would very much like your suggestion for the adoption, guidance or collaboration on this modeling. Establishing this will open up many wings and directions to consolidate in natural language grounding and as well CVPR use case oriented research for sure. Let's discuss. Regards, Saurav Karmakar ---------------------------- Dr. Saurav Karmakar Postdoctoral Researcher, Insight Centre, Data Science Institute, National University of Ireland Galway, Galway, Ireland
Received on Thursday, 9 July 2020 09:27:35 UTC