Re: Advancing Distributed Moderation

The information, feedback, concerns, calls for caution, and other points are all well-received here.

With respect to concerns regarding the consolidation of wealth, control, and power resulting from some comparative system designs, there are also open-source-community approaches underway (e.g., HuggingFace) which enable developers, teams, and companies to download AI models and subsequently utilize them – some for commercial purposes – on their own computing hardware. Locally-hosted LLMs can also allow greater control over practices affecting access to and ownership of data and this may make AI tools more appealing.

With respect to the computing resources for training new AI models, there is also a National Artificial Intelligence Research Resource (NAIRR) in development [1]. Work is presently underway on a pilot program [2].

As for public opinion and AI, here are some recent data:


  1.  52% of Americans are more concerned than excited about AI [3].
  2.  The first AI-related concern that I could find in the latest Chapman Survey of American Fears is #47 [4].

Synthesizing these data, it appears that these AI-related concerns are not pressing to the American people, but, when the topics are raised, e.g., in the forms and framings of opinion-polling surveys, the public is more concerned than excited.


Best regards,
Adam

P.S.: Towards better understanding the recent and unfolding national public dialog on AI, I am looking into reading some books next year [5][6][7][8]. I welcome any more recommendations on- or off-list.

[1] https://www.ai.gov/wp-content/uploads/2023/01/NAIRR-TF-Final-Report-2023.pdf

[2] https://new.nsf.gov/news/nsf-partners-kick-nairr-pilot-program


[3] https://www.pewresearch.org/short-reads/2023/11/21/what-the-data-says-about-americans-views-of-artificial-intelligence/

[4] https://www.chapman.edu/wilkinson/research-centers/babbie-center/_files/2023%20Fear/23csaf-9_high-to-low.pdf


[5] Cave, Stephen, and Kanta Dihal, eds. Imagining AI: How the World Sees Intelligent Machines. Oxford University Press, 2023.
[6] Bradford, Anu. Digital Empires: The Global Battle to Regulate Technology. Oxford University Press, 2023.
[7] Bader, Christopher D., Joseph O. Baker, L. Edward Day, and Ann Gordon. Fear Itself: The Causes and Consequences of Fear in America. (2020).
[8] Khan, Salman. Brave New Words: How AI Will Revolutionize Education (and Why That’s a Good Thing). New York, NY: Penguin, 2024.

________________________________
From: O'Brien, Sean <sean.obrien@yale.edu>
Sent: Saturday, November 25, 2023 5:52 PM
To: Emelia Smith <emelia@brandedcode.com>; James <jamesg@jamesg.blog>
Cc: Adam Sobieski <adamsobieski@hotmail.com>; public-swicg@w3.org <public-swicg@w3.org>
Subject: Re: Advancing Distributed Moderation

Thank you all for your thoughtful input in regard to moderation.

I would just like to add that there are (at least) four additional issues with AI/LLM in real-world scenarios:

* When we speak of "AI", we're talking about a handful of multi-billion-dollar backends owned / dominated by Big Tech intermediaries. Building solutions that require the usage of these systems puts entities like Microsoft, Alphabet, and Meta in an ever-more powerful position. In 1999 we were worried about Microsoft's monopolistic integration of IE in Windows. In 2006, the so-called "ASP loophole" was hotly debated during the GPLv3 drafting. In 2023, FOSS devs everywhere are inserting AI backends owned+operated by Big Tech into their software as essential, black-box middleware.

* AI/LLM has required the aforementioned piles of money to reach its current maturity, and that investment will eventually mean an end to freemium models or at least severe limitations on gratis usage, especially where API access and integration with FOSS projects is concerned. The infrastructure and energy requirements (read: costs) of OpenAI et al are rising rapidly, and any competitors attempting to meet similar scope and utility will carry huge financial and industrial baggage. All of this will place additional pressure on freemium revenue models. This is a prediction, of course, but nonetheless one borne from history and experience.

* We are starting to see real, tangible censorship of AI/LLM. That means there is default, upstream moderation before any possibility of downstream moderation, with little-to-no transparency from upstream.

* Any hypothetically ethical use of AI/LLM for moderation requires that users consent to their published works being used to train algorithms, be included in a training set, etc. In regard to FOSS licenses, I don't think I need to repeat the known issues with Microsoft's Copilot experiment for this list.

Cheers,
- Sean

--
Sean O'Brien

Lecturer, Yale Law School (Cybersecurity LAW 21314)
Fellow, Information Society Project at Yale Law School
Founder, Privacy Lab at Yale ISP, https://privacylab.yale.edu<https://privacylab.yale.edu/>




________________________________
From: Emelia Smith <emelia@brandedcode.com>
Sent: Saturday, November 25, 2023 5:04 PM
To: James <jamesg@jamesg.blog>
Cc: Adam Sobieski <adamsobieski@hotmail.com>; public-swicg@w3.org <public-swicg@w3.org>
Subject: Re: Advancing Distributed Moderation

I just wish to clarify, in my original reply, I said that whilst there could be applications for AI/LLMs/natural language models, the emphasis was that we need to be highly cautious and critical of such systems, given biases in training data and implementation.

That is, implementation of AI/ML without critically examining the systems will lead to further marginalisation and harm, NOT a reduction in harm.

There's been case after case of bad AI data, responses and faulty algorithms causing harm & chaos. Whatever steps you take around these technologies and applying them to the fediverse must be done with the utmost caution, and without assumptions of these AI models being inherently good.

There's a huge deal more that can be done in the moderation and trust & safety space before there's a _must_ for more advanced techniques. A good example of this is Mastodon's new android warnings with regards to replies: https://blog.joinmastodon.org/2023/11/improving-the-quality-of-conversations-on-mastodon/


Yours,
Emelia

On 25. Nov 2023, at 21:06, James <jamesg@jamesg.blog> wrote:


OpenAI has documented a way in which you can use GPT-4 for content moderation: https://openai.com/blog/using-gpt-4-for-content-moderation


I expect there to be a lot more work done in this field of research.

James

On Saturday, 25 November 2023 at 20:00, Adam Sobieski <adamsobieski@hotmail.com> wrote:

Emelia Smith,

Thank you for the feedback. I am new to the SWICG and am glad to see that there is some interest here in uses of AI/ML/CV/LLMs for equipping moderators and empowering end-users.


Best regards,
Adam

P.S.: I would enjoy learning more about the FIRES proposal.

________________________________
From: Emelia Smith <emelia@brandedcode.com>
Sent: Friday, November 24, 2023 12:24 PM
To: Adam Sobieski <adamsobieski@hotmail.com>
Cc: public-swicg@w3.org <public-swicg@w3.org>
Subject: Re: Advancing Distributed Moderation

Hi all,

Some of what Adam speaks of here is what I'm working on with my FIRES proposal, essentially a service for storing & sharing moderation advisories and recommendations, and looks past the current status quo of denylists or blocklists, allowing still for full federation restrictions, but also for more granular restrictions. It also allows for multiple different entity types, whether Domains/Instances, Hashtags, URLs, Users, etc.

When I worked at State in 2012/2013, we did structure opinions, where you'd have a topic and several words associated with that topic, and ultimately it didn't go mainstream.  So I'm skeptical on the "made me feel" proposal here, as it's additional "work" for the person using the service, maybe you could extrapolate this via emoji reactions, which are popular on the fediverse though?

Attempts at trying to classify content could help moderation, but it could also harm it, by reducing the "human touch", which creates problems like how Instagrams automated moderation often harms marginalised communities. So I do suggest caution in this path.

Additionally, this could then result in algorithmic timelines, which currently aren't present on the Fediverse. From what I've seen, there's generally a resistance to algorithmic timelines, even if those can help people who check social media less frequently than others.

You are right that AI/machine vision/LLMs could be beneficial for instance moderators, preventing exposure to harm, but we also need to be critical of and investigate the training of these models, and their human impact (there was a recent headline I read about OpenAI paying minimal amounts for human moderators to classify media as CSAM or not, which obviously has a human cost), additionally, we need to enquire as to the biases such algorithms may have and how they may adversely affect marginalised social groups.

Whilst tools can assist in moderation, ultimately I believe we need to have humans making the final decision.

(If anyone would like to peer-review FIRES I can send you an early draft, but I'm currently reworking a lot of it)

Yours,
Emelia Smith

On 24. Nov 2023, at 14:30, Adam Sobieski <adamsobieski@hotmail.com> wrote:


Social Web Incubator Community Group,

Introduction
Hello. I would like to share some ideas to better inform and equip distributed social media administrators and moderators.

Multimedia-content-scanning Rules and Service Providers
Software tools for moderators can be envisioned which:

  1.  Scan multimedia content on servers,
  2.  Provide moderators with time-critical information, updates, alerts, and alarms,
  3.  Provide moderators with real-time natural-language reports, data visualizations, and analytics.

Software tools for moderators could be updated in a manner resembling real-time "antivirus" data updates. Instead of there being one or a few such "antivirus" data providers, moderators could choose to subscribe to individual channels of data from multiple providers. As envisioned, each provider would serve updates across a number of described channels. Data sent in these channels, e.g., multimedia-content-scanning rules, could be merged on platform servers and subsequently utilized by moderators' software tools. Some data providers' channels might be free to use while others would require paid subscriptions.

As envisioned, moderators would be able to create, modify, and delete multimedia content scanning rules manually to customize their software tools.

Moderators' actions and decisions could be collected, aggregated, processed, and utilized for purposes including to distribute helpful real-time hints to other moderators across platforms. Moderators would be able to choose which service providers, if any, to share these data with. These data would not identify end-users or moderators.

The aforementioned "antivirus" data for informing and equipping decentralized social media moderators could additionally enhance and enable other technologies and systems, e.g., content-distribution algorithms and recommender systems.

With recent advancements to AI:

  1.
Systems like NeMo Guardrails<https://github.com/NVIDIA/NeMo-Guardrails> and Guardrails AI<https://github.com/guardrails-ai/guardrails> can enhance the capabilities of moderators' software tools,
  2.  Systems like GPT-4V<https://openai.com/research/gpt-4v-system-card> and LLaVA<https://github.com/haotian-liu/LLaVA> can process images occurring in social-media contexts in new ways.

Enabling More Granular Reactions to Content
Beyond liking content or not, end-users could react to content items by attaching text-keywords metadata. End-users could click on buttons next to content to enter open-ended text contents into text fields, e.g., "this content made me feel _____". There could be another button to use to add subsequent reactions – second, third, or fourth reactions – for more complex, and potentially mixed, reactions. Alternatively, these reactions could be comma-delimited, and converted to reaction tags as typed. End-users' reactions would tend to be drawn from a folksonomic vocabulary<https://en.wikipedia.org/wiki/Emotion_classification> and, accordingly, incremental search and recommendation features could reduce typing.

Anonymized usage data from end-users could be collected and sent to service providers. Envisioned data include anonymized status messages and anonymized reactions to content items. With these data, new usage trends, audience reaction data, natural-language reports, data visualizations, and analytics could be made available to moderators and administrators.

With unfolding advancements to AI, granular reactions to content could be explained and predicted. This could enhance tools for moderators such as automatically displaying warnings atop some content for end-users.

End-user-specified Preferences
Some moderators might want to allow their end-users to express content-related preferences, to allow end-users to directly or indirectly create, modify, and delete multimedia-content-scanning rules.

End-users' preferences can be complex; there would be some complex concepts or categories for machine-learning-based approaches to learn through the processing of items and end-users' responses to them. Consider, for instance, the following pseudocode for two rules, which end-users might express through a number of techniques, including natural language: "if an item is predicted to make me sad or angry, unless it is for activism or charity, I want a warning atop it."

Rule 1:

if(prediction(me, item, 'sad'))
{
    if(!(metadata(item, 'activism') || metadata(item, 'charity')))
    {
        return warning.content(item, 'sad');
    }
}
return warning.none;

Rule 2:

if(prediction(me, item, 'angry'))
{
    if(!(metadata(item, 'activism') || metadata(item, 'charity')))
    {
        return warning.content(item, 'angry');
    }
}
return warning.none;

Note that multiple warnings could be aggregated and displayed atop content items.

AI systems, i.e., large language models, could process end-users' natural-language expressions of content-related preferences into multimedia-content-scanning rules.

In theory, end-users' preferences could be collected, aggregated, and processed. These data would not identify end-users.

Conclusion
Existing and new standards and recommendations can enable indicated technology scenarios for better informing and equipping distributed social media administrators and moderators.

It would be great if, after some amount of delay, these kinds of data, indicated above, could be made available to the scientific community and to the public.

Thank you. Any comments, questions, or thoughts on these ideas?


Best regards,
Adam Sobieski

P.S.: See also:

  *   https://github.com/swicg/general/issues/34

  *   https://github.com/swicg/general/issues/7

  *   https://github.com/w3c/activitypub/issues/231

  *   https://github.com/w3c/activitypub/issues/232




<publickey - jamesg@jamesg.blog - 0xC06B40B5.asc>

Received on Monday, 27 November 2023 10:22:08 UTC