W3C home > Mailing lists > Public > semantic-web@w3.org > March 2019

Re: Solomon''s curse and search Bias

From: ProjectParadigm-ICT-Program <metadataportals@yahoo.com>
Date: Mon, 4 Mar 2019 19:18:35 +0000 (UTC)
To: Thomas Passin <tpassin@tompassin.net>, Paola Di Maio <paoladimaio10@gmail.com>, Sherman Monroe <sdmonroe@gmail.com>
Cc: "schema.org Mailing List" <public-schemaorg@w3.org>, SW-forum <semantic-web@w3.org>, "public-aikr@w3.org" <public-aikr@w3.org>, Kingsley Idehen <kidehen@openlinksw.com>, Hugh Williams <hwilliams@openlinksw.com>, vios <vios@dev-team.com>
Message-ID: <190194130.9154823.1551727115200@mail.yahoo.com>
Here is my  take based on the same problem I encounter every day when searching for information that deals with sustainable development AND other issues.
Western search engines like Google and Bing and Yahoo Search use algorithms like PageRank, which use a one size fits all approach as a starting point and then use user search history and location to adapt both search results and advertising to more or less suit the user's needs.
There are several approaches possible to search.
(1) the approach used by Google and its western competitors(2) the approach used by Chinese search engines which shamelessly and without any concern for privacy and digital rights uses deep learning and massive amounts of data and data gleaned thru webcam cameras from the eye movements of users to create user profiles that center on geographical location and demographics to suit user needs, mainly for online shopping and other established prioritized user activity when browsing and searching(3) a brute force approach which somehow has to organize all the results into categories, e.g. the librarian category folders, regardless of user location and demographics, e.g. as promoted by the General Data Protection Regulation of the European Union(4) a brute force approach with category folders but with user location and demographics, which we are able to input to an input form similar to those for acceptance of cookies(5) an AI approach which uses a form to let the user allow for user location and demographics use and creates a popup dashboard which we can continuously tweak and modify. This has several advantages. The first and foremost to let the user allow the AI to heuristically prune the research results and also offers the user to see both the results with or without pruning. This approach will by default use category folders.
The last three approaches will substantially reduce the need for deep learning and massive amounts of data gathered without consensus or worse, without  permission, are in compliance with the GDPR and create a consensus based interactive AI approach in online search.
In my humble opinion options 4 and 5 are the ways to go. Milton Ponson
GSM: +297 747 8280
PO Box 1154, Oranjestad
Aruba, Dutch Caribbean
Project Paradigm: Bringing the ICT tools for sustainable development to all stakeholders worldwide through collaborative research on applied mathematics, advanced modeling, software and standards development 

    On Monday, March 4, 2019 10:01 AM, Thomas Passin <tpassin@tompassin.net> wrote:
 

 On 3/4/2019 6:54 AM, Paola Di Maio wrote:
> Sherman and Thomas P
> thanks a lot for sharing your search journey with friendly narrative and
> about this project. Looks really neat!
> 
> I am still thinking about the need to carry ot some structured data 
> search on open web via  general search engines tho.  I hope Google may 
> consider adopting your architecture or at least some of these ideas 
> which in principe work at least within a closed data set,

The kinds of things you are trying to explore can also be approached 
using a framework of library science, and specifically the concepts of 
"navigation" and "collocation".  Collocation means that items that are 
somehow similar can be found "near" each other in some sense of the word 
"near".

For a look an an attempt to provide good navigation and collocation for 
a small private data set, you could look at my paper "Browser bookmark 
management with Topic Maps" for the Extreme Markup Languages conferences 
from 2003 -

http://conferences.idealliance.org/extreme/html/2003/Passin01/EML2003Passin01.html

This work was my effort to get the most out of very limited amount of 
information, namely, titles of web pages. The viewpoint behind the work 
derived from the library science concepts of collocation, navigation, 
and "subject language".  To quote from my paper: "A subject language is 
a vocabulary for describing the subject of a work in such a way that it 
can be recognized and thereby found — that is, to provide navigation 
capability for a collection."

Another point to keep in mind is that a good user interface must be very 
different for a small project as contrasted with one that encompasses a 
large collection of data.  Just for a tiny example, a pick list for ten 
items can be usable whereas one for 10,000 items is not.  A UI for a 
large data set is hard to design even if your system has good 
collocation and navigation properties.

TomP





   
Received on Monday, 4 March 2019 19:24:01 UTC

This archive was generated by hypermail 2.3.1 : Monday, 4 March 2019 19:24:02 UTC