- From: Owen Ambur <Owen.Ambur@verizon.net>
- Date: Thu, 26 Mar 2020 23:55:06 -0400
- To: public-aikr@w3.org
- Message-ID: <bf1aed6e-b891-4cc2-4dc6-4e969403be96@verizon.net>
The article says of "Harms of representation": It gets tricky when it comes to systems that represent society but don’t allocate resources. These are representational harms. When systems reinforce the subordination of certain groups along the lines of identity like race, class, gender etc. It is a long-term process that affects attitudes and beliefs. It is harder to formalize and track. It is a diffused depiction of humans and society. It is at the root of all of the other forms of allocative harm. Representational harm is impossible to track without formalizing and recording metrics for the impact upon the relevant stakeholder groups. Moreover, to be humanly comprehensible at "Big Data" scale, such metrics must be gathered and shared in an open, standard, machine-readable format, like StratML Part 2. In any event, the objectives of Microsoft's FATE group are now among seven MS's plans available in StratML format at https://stratml.us/drybridge/index.htm#MS or, more specifically, https://stratml.us/carmel/iso/FATEwStyle.xml It would be good if the FATE folks could demonstrate leadership by example in reporting their plans and results in machine-readable format. Hopefully, they must be collaborating with MS's Project Cortex group but it would be good if they were to use a data structure like the stratml:Relationship elements <https://stratml.us/references/oxygen/PerformancePlanOrReport20160216_xsd.htm#Relationship> to make the logical connections salient and report progress. https://stratml.us/carmel/iso/MSPCwStyle.xml BTW, as far as government and politics are concerned, I believe we can do better than dictatorial majoritarian "representation" -- which, by definition, discriminates against minorities and also presents false choices for many, if not most purposes. https://www.linkedin.com/pulse/transforming-governance-reducing-cost-gofpau-owen-ambur/ Owen On 3/26/2020 10:45 PM, Paola Di Maio wrote: > I share this interesting article > https://hub.packtpub.com/20-lessons-bias-machine-learning-systems-nips-2017/ > > > In particular, emphasis on 'representational harm' which I think > should be imperative > we address in our work > > I ll enter this in Zotero > > PDM
Received on Friday, 27 March 2020 03:55:19 UTC