- From: Timothy Holborn <timothy.holborn@gmail.com>
- Date: Sun, 9 Apr 2023 01:11:07 +1000
- To: Mark Lizar <mark@openconsent.com>
- Cc: Adam Sobieski <adamsobieski@hotmail.com>, Mark Hampton <mark.hampton@ieee.org>, "public-humancentricai@w3.org" <public-humancentricai@w3.org>
- Message-ID: <CAM1Sok20j4i+qRco8C7Zb_iDpAJeYYF6omo+fuFnQL8+krBfQA@mail.gmail.com>
Mark, Its fantastic to have you involved, and i look forward to getting stuck into the process of better understanding your work, which, imo, appears remarkable.... will follow-up l8r. but, fwiw. ack. Timothy Holborn. On Sun, 9 Apr 2023 at 00:51, Mark Lizar <mark@openconsent.com> wrote: > Greetings HCAI, > > It is nice to be apart of this group, thank you Tim for the invitation. > It is amazing to see this topic, finally break through into a standards > group, and amazing to see the rapid growth in awareness of this space in > the last couple of months. Truly showing the potential of ai for bad or > for planet saving goodness. > > I believe digital equality should be inclusive and available to all and > that this is required so that people can control their own data, where ever > it may be. A critical digital freedom in my opinion, > > Which is why I propose a focus on Digital Transparency standard, that is > used to generate notice records and consent notice receipts, for every > digital processing interaction, through any interaction with notice, > notifications, disclosures, and even physical signs. > > This implementation approach, is one that proposes that a mirrored record > of processing activities is generated from a notice (rather than tick box > policy agreements) in which people get what is essentially, a receipt. A > receipt, which (FYI) is likely the oldest human form of human writing > <https://numismatics.org/pocketchange/receipts/>and the first ever form > of scripting. The receipt solves for the issue of trust, and implemented > to capture meta-data includes everyone. equality in records. > > My work has championed consent standards over the last decade as well as > the Consent Receipt information structure and technology. This has been > developed out of research as a social scientist and as digital identity > engineer, I champion the consent and notice receipts @ the Kantara > Initiative. Since the volunteer work began in 2013, it has successfully > been included in ISO/IEC 29184 Online privacy notice and consent standard > and is the basis for ISO/IEC 27560 - consent record information structure. > (In progress). Which means it is a means that can implement the solution > architecture for human centric AI. > > To this point, as humans, consent is how we control data, and information > in context, it is not an out of context tool (for us). Systems, manage > permissions, humans manage consent, which means we manage many permissions > in context, contrary to what tech has us all believing. Clarity in > semantics, and the ontologies involved have been the battle ground for > these concepts. As such it is remarkable we do not have our own records > of digital relationships, so that we can govern our own data and manage > AI. We needs these records to start the arms race. > > Digital Notice records are, and have been for a long time required, but > not produced by industry. We came close in 2015, but, now we have a > better alternative at the Kantara Initiative ANCR WG > <https://kantara.atlassian.net/wiki/spaces/WA/pages/164823104/ANCR+Digital+iDentity+Transparency+Framework+DT-Levels+of+Assurance> - > with a human trust framework for making records and receipts that > supercedes T&C’s. > > ANCR - Digital Transparency Framework > <https://kantara.atlassian.net/wiki/spaces/WA/pages/164823104/ANCR+Digital+iDentity+Transparency+Framework+DT-Levels+of+Assurance>, > which uses an international adequacy and transparency baseline, has > transparency performance indicator’s > <https://kantara.atlassian.net/wiki/spaces/WA/pages/82542593/Trust+Performance+Indicator+s+TPI+s>, > which generate PII Controller Credentials > <https://kantara.atlassian.net/wiki/spaces/WA/pages/114098237/Open+Notice+Record+PII+Controller+Credential>(digital > identity) that defined for public interest and inherently enable privacy > Ai. > > Imagine what we could do if we had our own records, and that every data > processing activity produced a receipt which was added to these records. > We wouldn’t need to ask for our data to be deleted anymore, we could do > that with multiple providers, in context with the click of one button. We > would have personalized digital privacy policies, not privacy policies for > contracts (T&C’s) > > With a consent based authorization protocol, (like AuthC > <https://kantara.atlassian.net/wiki/spaces/WA/pages/38174721/AuthC+Protocol>) > we wouldn’t need logins for every service either. We could use verified > credentials validated by regulated 3rd party governing authorities, > > For high assurance active privacy state monitoring, using our own records > would help scale trust online (the main use of privacy ai), Every > interaction we have with any AI would then feeds our own, private ai, our > personal micro-data, and our more intelligent data could then be > pernissioned according to a common set of rules, (digital privacy Magna > Carta) and together we could become a smart species. (Thats the vision > anyway) > > Which is why I urge a starting point in which a Digital Privacy AI is > implemented as a Co-Regulatory framework, so that we can stop making > everyone fill out there personal information for every service everywhere. > To start a network protocol where processing of personal data requires a > Controller credential so there are no unknown 3rd parties. Then we control > our own data sources. This promising approach to regulate AI, also > addresses mis-information at scale, and provide the foundation framework to > govern future tech, like quantum and government AI’s. > > This year in 2023, we finally have the laws, standards, and with the Jan 4 > decision pro-consent against Meta, the legal precedence to implement global > market strength digital transparency, digital rights and data control, > > Nice to meet you all, > > Mark Lizar > > PS - I have a W3C Do Track - Digital Consent Use Case for Private AI, if > anyone is interested? The idea is to cut out the commercial intermediaries > like google and Facebook, but with our own Privacy AI. An AI that uses > interactional privacy law to leverage the data we have in these services, > for our own empowering services. (Independent of big tech) > > > > > > > > > > On Apr 8, 2023, at 9:23 AM, Adam Sobieski <adamsobieski@hotmail.com> > wrote: > > Mark, > > Thank you for the useful hyperlink about differentiable technological > development. > > I see your points about personalization, preferences, and configurability. It > seems that we can envision something like a configurable *advanced > Grammarly* which would be equally available to all participants, > councilmembers, representatives, legislators, and Congresspeople. Perhaps > the composition-related functionalities in Web browsers' AI-enhanced > sidebars may evolve into something similar. > > When I consider lawyers in courtrooms and participants in group > deliberations (e.g., legislators) what comes to my mind, in this particular > discussion, is that I hope that they are all, or that their staffs are all, > equally technologically proficient. Imagine a courtroom where one party’s > lawyer had a laptop connected to a suite of advanced tools via Wi-Fi > Internet while the other lawyer had a briefcase with a stack of papers. > > Judicial and legislative systems seem to be more intuitively fair when all > of the participants are equally proficient and equally equipped with > preparation-related and performance-related tools. There could even be > "arms race" scenarios in terms of speechwriting assistants, debate coaches, > and other preparation-enhancing and performance-enhancing tools. > > Arguments for *advanced forum software* (e.g., features delivered via > plugins for social media software) include, but are not limited to, that: > > > 1. Some of these tools should obtain and process discussion contexts > to provide better composition-related advice to each user. > 1. However, per-user tools like *advanced Grammarly* or *Web > browser sidebar composition tool* could, in theory, scrape the > entireties of discussion threads, from webpages, up to the posts being > authored (perhaps utilizing HTML5 document metadata, Web schema, or linked > data) to deliver better advice requiring discussion contexts. > 2. Some of the tools under discussion are useful for evaluating the > posts of other users or evaluating combinations of posts from multiple > users, entire threads, as they unfold. > > > > Best regards, > > Adam > > ------------------------------ > *From:* Mark Hampton <mark.hampton@ieee.org> > *Sent:* Saturday, April 8, 2023 7:14 AM > *To:* Adam Sobieski <adamsobieski@hotmail.com> > *Cc:* public-humancentricai@w3.org <public-humancentricai@w3.org> > *Subject:* Re: Artificial Intelligence and Group Deliberation > > Hi Adam, > > I imagine those tools would need to be personalized (I'd prefer owned by > the user), one person's propaganda is another's preference. There would be > broader dynamics that could be spotted through collaboration of those > personalized systems sharing information or using shared sources of > information. > > There is an almost certain risk of information overload and that seems to > make people more manipulable. If human centric means caring for humans then > I think we need to be careful. Human centric AI could become a way of > accelerating AI rather than caring for humans - I would really like to see > human centric AI leading to > https://en.wikipedia.org/wiki/Differential_technological_development rather > than mitigations for inhuman technologies. > > The current direction of technological progress does not seem very human > centric at all. The work to build technical solutions to problems > introduced by technology seems to be a symptom of this. I don't see any > current very important short/medium term human material problems that need > AI but I'm open to being convinced otherwise. Technologists (and I speak as > one) risk to have a hard time accepting they are part of the problem rather > than the solution. > > An off the cuff reaction but I hope it is of some use to you. > > Kind regards, > Mark > > > On Sat, Apr 8, 2023 at 12:57 AM Adam Sobieski <adamsobieski@hotmail.com> > wrote: > > Human-centric AI Community Group, > > Something that Timothy Holborn said in a recent letter to this mailing > list reminded me of some thoughts that I had about AI a few years ago. At > that time, I was considering uses of AI technology for supporting > city-scale e-democracies and e-townhalls. I collated a preliminary > non-exhaustive list of tasks that AI could perform to enhance public > discussion forums: > > 1. Performing fact-checking > 2. Performing argument analysis > 3. Detecting spin, persuasion, and manipulation > 4. Performing sentiment analysis > 5. Detecting frame building and frame setting > 6. Detecting agenda building and agenda setting > 7. Detecting various sociolinguistic, social semiotic, sociocultural > and memetic events > 8. Detecting the dynamics of the attention of individuals, groups and > the public > 9. Detecting occurrences of cognitive biases in individual and group > decision-making processes > > With respect to point 3, a worry is that some participants in a community > might make use of AI tools to amplify the rhetoric used to convey their > points of view. These were concerns about technologies like: "virtual > speechwriting assistant" and "virtual debate coach". > > Some participants of an e-townhall or social media forum might make use of > AI tools to spin, to persuade, to manipulate the other members for their > own reasons or interests or might do so on behalf of other parties who > would pay them. > > My thoughts were that technologies could mitigate these technological > concerns. Technologies could monitor large-scale group discussions, on > behalf of the participants, while serving as tools available to all of the > participants. For example, AI could warn content posters before they posted > contentious content (contentious per their agreed-upon rules) and > subsequently place visible icons on contentious posts, e.g., content > detected to contain spin, persuasion, or manipulation. > > I was brainstorming about solutions where AI systems could enhance group > deliberation, could serve all of the participants simultaneously and in an > open and transparent manner, and could ensure that reason prevailed from > group discussions and deliberations. Today, with tools like GPT-4, some > of these thoughts about humans and AI systems interoperating in public > forums, e-townhall forums and social media, seem to be once again relevant. > Any thoughts on these topics? > > > Best regards, > Adam Sobieski > > >
Received on Saturday, 8 April 2023 15:11:52 UTC