- From: Kim Hamilton <kimdhamilton@gmail.com>
- Date: Sat, 18 Apr 2026 15:05:46 -0700
- To: Adrian Gropper <agropper@healthurl.com>
- Cc: "Michael Herman (Trusted Digital Web)" <mwherman@parallelspace.net>, Christoph <christoph@christophdorn.com>, Melvin Carvalho <melvincarvalho@gmail.com>, "Eduardo C." <e.chongkan@gmail.com>, Marcus Engvall <marcus@engvall.email>, W3C Credentials CG <public-credentials@w3.org>, "Manu Sporny (msporny@digitalbazaar.com)" <msporny@digitalbazaar.com>
- Message-ID: <CAFmmOzdaEP_i6Y3Pr8yYDLLzMx_BtHis77nfzxmVrYkAurEddA@mail.gmail.com>
Replace the word “fair” with “relevant “. I’m not weighing in on the fairness, just the relevance to this group. On Sat, Apr 18, 2026 at 2:46 PM Kim Hamilton <kimdhamilton@gmail.com> wrote: > If you think the role/goal of this group is code generation, that's a fair > argument. > > On Sat, Apr 18, 2026 at 2:42 PM Adrian Gropper <agropper@healthurl.com> > wrote: > >> The structure of workgroups and SDOs will change with increasingly >> capable AI. >> >> For almost a year now, I've been able to ask LLMs how to accomplish what >> I want. They read the API documentation, answered my questions and >> implemented the API. I never, once looked at a standard or the code. >> However, I do ask Claude Code to generate markdown files for documentation, >> which helps me feel in control and reduces the cost and risk of future >> changes. The documentation includes very valuable analyses of security >> vulnerabilities. >> >> Consequently, the role of groups like CCG and standards workgroups is >> changing, at least for me. I look forward to learning about business >> realities and real-world experience. >> >> I no longer care about new standards. If a vendor or service provider >> wants my business, it's up to them to provide and document the APIs. >> Standardized APIs can reduce risk and switching costs, of course, but if >> the tradeoff is 5+ years of discussions on CCG and related forums, the >> juice is no longer worth the squeeze. >> >> Adrian >> >> On Sat, Apr 18, 2026 at 4:22 PM Kim Hamilton <kimdhamilton@gmail.com> >> wrote: >> >>> Marcus represents a growing contingent of increasingly concerned folks, >>> and I'm grateful to him for speaking up. >>> >>> Several factors: >>> 1. LLMs trained on the existing corpus of internet standards will tend >>> to reproduce the assumptions baked into it. Particularly given the work of >>> this group, our experience* as humans *is critical for detecting / >>> evaluating any such bias*. >>> 2. Particularly for a standards generation/incubation body, it's >>> essential to know that a contribution comes from an individual/org/entity >>> with a stake in the outcome and accountability for the direction it implies. >>> 3. Of particular concern is (what I consider) a category error >>> prematurely attributing properties like "knowledge", "understanding" to >>> LLMs, with accompanying statements implying we humans are now off the >>> hook for deeper critical evaluation. >>> >>> Our human agency, judgment, and accountability*,* *feeble though our >>> little brains may be,* are needed now more than ever. >>> >>> To Manu's point, perhaps the venue for authentic human discourse in the >>> CCG is now restricted to the group calls, until someone creates a >>> vocally-convincing agent... >>> >>> Kim >>> >>> ** Not saying you cannot use an LLM to help with this work, but see >>> other points* >>> >>> On Sat, Apr 18, 2026 at 12:38 PM Michael Herman (Trusted Digital Web) < >>> mwherman@parallelspace.net> wrote: >>> >>>> [I know I’m saying too much but I have a lot to say.] >>>> >>>> >>>> >>>> RE: LLMs will slowly absorb all pattern matching and synthesis work. >>>> >>>> >>>> >>>> +1 Christoph >>>> >>>> >>>> >>>> At DAVOS this past January, noted expert Yuval Noah Harari shared the >>>> following (https://youtu.be/QiT2yK-5-yg?si=x71xvnZou_72o9u2&t=227): >>>> >>>> >>>> >>>> *Some people argue that AI is just glorified autocomplete. It barely >>>> predicts the next word in a sentence. But is that so different from what >>>> the human mind is doing? Try to observe - to catch - the next word that >>>> pops up in your mind. Do you really know why you thought of that word? >>>> …where did it come from? Why did you think of this particular word and not >>>> some other word? Do you know? * >>>> >>>> >>>> >>>> *As far as putting words in order is concerned, AI already thinks >>>> better than many of us. Therefore, anything made of words will be taken >>>> over by AI. If laws are made of words, then AI will take over the legal >>>> system. If books are just combinations of words, then AI will take over >>>> books. * >>>> >>>> >>>> >>>> *If religion is built from words, then AI will take over religion. This >>>> is particularly true of religions based on books like Islam, Christianity, >>>> or Judaism. Judaism called itself the religion of the book and it grants >>>> ultimate authority not to humans but to words in books. Humans have >>>> authority in Judaism not because of our experiences but only because we >>>> learn words in books. Now, no human can read and remember all the words in >>>> all the Jewish books. But AI can easily do that. What happens to a >>>> “religion of the book” when the greatest expert on the holy book is an AI?* >>>> >>>> [Yuval Noah Harari, 2026] >>>> >>>> >>>> >>>> *In effect, all the standards we need already exist. It’s a simple >>>> matter of choosing the right words and placing them in the right order. AI >>>> can do this more completely, more correctly, with greater precision, and >>>> less time and effort than any human or (working) group of humans. [Michael >>>> Herman, 2026]* >>>> >>>> >>>> >>>> I’ll try to pipe down, >>>> >>>> Michael Herman >>>> >>>> Chief Digital Officer >>>> >>>> Web 7.0 Foundation >>>> >>>> >>>> >>>> *From:* Christoph <christoph@christophdorn.com> >>>> *Sent:* Saturday, April 18, 2026 8:23 AM >>>> *To:* Melvin Carvalho <melvincarvalho@gmail.com>; Eduardo C. < >>>> e.chongkan@gmail.com> >>>> *Cc:* Marcus Engvall <marcus@engvall.email>; W3C Credentials CG < >>>> public-credentials@w3.org> >>>> *Subject:* Re: The Slopification of the CCG >>>> >>>> >>>> >>>> On Sat, Apr 18, 2026, at 2:04 AM, Melvin Carvalho wrote: >>>> >>>> so 18. 4. 2026 v 5:39 odesílatel Eduardo C. <e.chongkan@gmail.com> >>>> napsal: >>>> >>>> "I find it difficult to trust a contribution in this group if it has >>>> been generated by an LLM" >>>> >>>> >>>> >>>> A- I wonder how everyone can tell if something was written by an LLM? >>>> Aside of the now infamous "--" here and there that it uses, how can you >>>> guys tell? ( how do you know it is not a grammaraly plugin? ) >>>> >>>> B- Also wondering if the embedded gemini would detect if an email or >>>> text was generated by an LLM. And more importantly, detect slope in that >>>> email or content. e.g. I normally use 2 different LLMs to do manual >>>> adversary checks on each other outputs and analysis, Gemini + Claude, and >>>> they usually find improvements or catches, and I also find deviations and >>>> correct the alignment. >>>> >>>> C- Most slope happens when one is researching or asking for things that >>>> are not in the model itself. E.g. you ask for certain uncommon thing and >>>> the models—all of them—keep levitating and pointing towards what the >>>> probability says they should answer. One needs to be aware of that. >>>> >>>> >>>> >>>> LLM content is reasonably easy to identify as many signals are inserted >>>> by default. >>>> >>>> >>>> >>>> If we consider content on the internet over the last 2-3 years its gone >>>> from small LLM contributions to majority LLM content. >>>> >>>> >>>> >>>> I see the same happening with standards, as LLMs get smarter. >>>> >>>> >>>> >>>> IMHO we're in the last phase of human authored standards, and LLMs will >>>> end up becoming the majority of content in standards. But that's nothing to >>>> fear. It just means we get things over the line faster and at a higher >>>> quality than ever before. >>>> >>>> >>>> >>>> The standards that went before will be building blocks for what comes >>>> next. >>>> >>>> >>>> >>>> I share this point of view. LLMs will slowly absorb all pattern >>>> matching and synthesis work. >>>> >>>> >>>> >>>> There is currently a huge spectrum of LLM practitioner competence >>>> leading to different assessments of LLM usefulness and capability. This >>>> leads to opinions that are not well grounded. >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> What humans will be able to do is to manage the complexity budget, >>>> present use cases and help standards work gain adoption. >>>> >>>> >>>> >>>> Humans will provide the judgement to direct LLMs and thus standards >>>> towards what matters which is what LLM cannot do. >>>> >>>> >>>> >>>> Humans provide direction, LLMs execute. >>>> >>>> >>>> >>>> IMO the focus of this list will evolve towards making such judgements >>>> and discussions about meaningful direction will become more and more >>>> important in the future. When you can go any direction rapidly, you might >>>> as well go in directions that really matter. >>>> >>>> >>>> >>>> Christooh >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> BTW, I agree with Michael Herman 100%. >>>> >>>> -- >>>> >>>> Eduardo Chongkan >>>> >>>> >>>> >>>> >>>> >>>> On Fri, Apr 17, 2026 at 4:43 PM Marcus Engvall <marcus@engvall.email> >>>> wrote: >>>> >>>> Hi all, >>>> >>>> >>>> >>>> I have been a passive observer of the CCG and have found the >>>> discussions in this group to have been remarkably considered, professional, >>>> and above all else clear in both intent and direction. I hesitate to >>>> comment on the current state of the mailing list as my tenure is minuscule >>>> compared to some of my brilliant co-participants, but the quality of recent >>>> contributions have compelled me to share some thoughts. >>>> >>>> >>>> >>>> Standards work is fundamentally a rigorous process of deriving a >>>> synthesis of human knowledge and judgement through healthy debate and, >>>> particularly in this group, decentralised knowledge discovery. It is >>>> precisely the provenance of consideration that establishes the trust basis >>>> necessary for the voluntary adoption of standards. Without trust, there is >>>> no standard. It follows then that preserving the integrity of the >>>> standardisation process is existential for any group working on standards. >>>> >>>> >>>> >>>> AI has improved the accessibility of standardisation to a larger and >>>> more diverse group of participants which is incredibly valuable for >>>> standardisation and should be encouraged. However, it should not come at >>>> the cost of compromising the integrity of the process itself, something I >>>> fear is happening in this group. >>>> >>>> >>>> >>>> Many recent contributions on this mailing list bear the hallmarks of >>>> LLM generation. To be clear, it is my view that there is nothing wrong with >>>> using AI agents to assist with research, proofreading, and other similar >>>> tasks. I use these tools every day professionally and their value is >>>> undeniable. That said, they are not replacements for human judgement, and >>>> this is something I think shared by most people in this group. >>>> >>>> >>>> >>>> I find it difficult to trust a contribution in this group if it has >>>> been generated by an LLM, and it is becoming increasingly intractable to >>>> follow discussions as they seem to inevitably degenerate to chatbots >>>> arguing with each other. Inferring the direction of standardisation, which >>>> has a direct impact on commercial and technical planning, becomes >>>> impossible. I find it quite ironic that the recent thread discussing LLMs >>>> and agents in the CCG contains responses that suggest that they themselves >>>> have been generated by an AI. If anything, I think it is proof enough of >>>> how acute this problem is. >>>> >>>> >>>> >>>> There is also the somewhat primal and adversarial aspect of evaluating >>>> human judgement and reaching consensus. A debate is a contest between two >>>> humans arguing for their position, which presupposes real agency and, well, >>>> humanity. An AI agent is not, and will never be, a real human - and nobody >>>> wants to credibly evaluate the arguments of a robot. >>>> >>>> >>>> >>>> I am not sure what the solution is, but I feel that the effects of this >>>> are severe and will almost certainly discourage participants from >>>> contributing, the downstream consequences of which I think are clear to >>>> everyone. >>>> >>>> >>>> >>>> I would like to close out this lengthy email with this: I think a >>>> serious discussion should be opened to consider migrating to a discussion >>>> channel that is more resistant to AI agents, or at least consensus be >>>> formed to institute and enforce a strict code of conduct with >>>> zero-tolerance for AI slop. Openness is important, and exclusionary >>>> dynamics must be avoided to the extent possible, but the integrity of the >>>> standardisation process and the important work done in this group depends >>>> on humanity and not artificiality. >>>> >>>> >>>> >>>> Sincerely, >>>> >>>> >>>> >>>> -- >>>> >>>> Marcus Engvall >>>> >>>> >>>> >>>> Principal—M. Engvall & Co. >>>> >>>> mengvall.com >>>> >>>> >>>> >>>
Received on Saturday, 18 April 2026 22:06:02 UTC