Re: The Slopification of the CCG

Not to denigrate anyone's posts, including this one, but one complaint
seems to be the length of posts contributed by AI.  I find it ironic that
many such posts are quite long.  I personally find that content
interesting, but perhaps, in the interest of people with less time to read
than I, any post longer than a paragraph or two should start with a tldr.

--------------
Alan Karp


On Thu, Apr 23, 2026 at 10:20 AM Moses Ma <moses.ma@futurelabconsulting.com>
wrote:

> Nikos,
>
> I wanted to applaud your candor, which was refreshingly straightforward.
> However, I urge everyone to adhere to the unspoken goal of inclusivity
> here. We need to be peacemakers.
>
> Anyway, your post primed the pump of ideas.
>
>  1) A small experiment I ran recently showed ~25% of people in this forum
> can’t reliably distinguish LLM output from human writing. So some of what
> gets labeled “AI slop” is actually just perception. But that’s almost
> beside the point here. The real issue isn’t AI—it’s verbosity as a
> strategy.
>
> 2) Standards groups generally follow a few structural dynamics:
>
>    - inclusion > exclusion
>    - visibility = influence
>    - language = control
>
> What’s changed is that AI has reduced the cost of writing. So individuals
> who were already inclined to, ah, “over-contribute” can now scale that
> behavior, flooding the channel with low-signal, self-promotional, or
> tangential content. These people are like the blowhards at a company who
> believe that talking really loud and constantly mansplaining is a success
> strategy. This is not the kind of leadership we need in the 21st century.
>
> The failure mode isn’t just annoyance—it’s attention capture by volume,
> where a few verbose participants degrade signal-to-noise for everyone else.
> What we really need is an AI with a  *bloviation sensor*. Most of us do
> this internally by simply not reading certain posts, until we’ve had enough
> and lash out. Then the bloviator is justified in feeling attacked.
>
> Therefore, instead of  debating tools or reputation, it may be more
> productive to consider the possibility of placing lightweight guardrails on
> contribution quality:
>
>    - Contribution caps per cycle (forces prioritization)
>    - One idea per message (no multi-topic dumps or longwinded responses
>    like this one)
>    - Editorial compression rights (chairs or AI can edit and summarize
>    without loss of weight)
>    - Track signal-to-noise to reward reputation for increased caps (what
>    actually survives into the draft)
>
> Something like this would get us out of policing individuals or tools, and
> back to protecting the quality of the work.
>
> 3) I think that in a few years, people who refuse to use LLMs—essentially
> “artisanal writers”—will seem as quaint as luddites who refuse to use
> Google.   I added the em dashes to show that a human generated response can
> still use em dashes, they are not the six fingers of LLM writing. I like
> them because they force the mind to “slow the breath” while reading.
>
> To wrap up, the happy ending I’d love to see is something likely
> impossible. My preference is a new kind of process that nurtures growth, by
> encouraging the hesitant to find their voice, novices to get up to speed
> faster, and the emergence of greater self-awareness by the bloviators.
>
> I’m actually working on a Web 89.0 vision of this vision (haha)….
>
> My incubator is building a “stealth-ish mode” startup called EmergentYOU
> with the goal of creating something that could provide a labor transition
> cushion for the AI era—using longitudinal coaching, hyperpersonalized
> learning pathways, and an AI career co-pilot to continuously align people
> with opportunity. It converts disruption into mobility by linking skills,
> employers, and outcomes in a closed-loop system that compounds human
> potential over time. We’ll announce the EmergentYOU concept at Human Tech
> Week in San Francisco next month.
>
> Anyway, I’ve been thinking about we’re extending EmergentYOU into
> EmergentUS, to support teams: an intelligence layer designed to increase group
> cohesion, reduce participation disparity, and enhance group consonance.
> The system nudges quieter participants to contribute, modulates dominant
> voices, and elevates high-signal input—creating balanced, adaptive dialogue
> and measurably stronger collective performance. However, the real issue is
> that the entire team would need to agree to undergo the process. If there
> is interest in piloting EmergentUS in an SDO context… let me know.
>
> – Moses
>
>
> PS, if you’re in the SF Bay Area and would like to attend our event at
> Human Tech Week… let me know too.
>
>
> <moses@nureon-eda.ai>
>
> On Apr 23, 2026 at 12:55 AM, <NIKOLAOS FOTIOY <fotiou@aueb.gr>> wrote:
>
> Hi all,
> I think we are losing the context here. The problem is that certain
> perfectly identifiable individuals spam the list with mostly meaningless,
> self-promotional content. For example, every couple of messages I receive
> some web 7.0 irrelevant, non sense. AI tools have just made their job
> easier to generate content. Blaming AI tools is just a polite way to say to
> those individuals “please stop you are creating too much noise”
>
> Best,
> Nikos
>
> 23 Απρ 2026, 10:36 πμ, ο χρήστης «Amir Hameed <amsaalegal@gmail.com>»
> έγραψε:
>
> 
> Hi Adrian,
>
> I do not think the concern is about restricting the use of tools. People
> will use whatever tools are available to them—that’s inevitable.
>
> The issue is that reputation alone is not a strong enough primitive for
> systems that aim to operate at scale and across jurisdictions.
>
> In distributed environments, we typically rely on properties like:
>
>    - verifiable provenance
>    - non-repudiation
>    - integrity of authorship
>
> These are not about limiting expression, but about ensuring that
> contributions can be evaluated independent of the individual’s perceived
> credibility.
>
> Saying “my reputation will suffer if I’m wrong” assumes:
>
>    1. reputations are consistently observable across contexts, and
>    2. reputational consequences are sufficient deterrent
>
> In practice, neither assumption holds reliably—especially in global,
> asynchronous systems.
>
> On enforcement: Global enforcement is not realistic. That’s precisely why
> systems tend to push guarantees down to verifiable layers rather than
> relying on behavioral expectations at the application layer.
>
> So perhaps the problem is not tool usage vs. responsibility, but:
>
> how do we make authorship and intent more verifiable without constraining
> participation?
>
> Regards
> Amir Hameed
>
> On Thu, 23 Apr 2026 at 12:15 PM, Adrian Gropper <agropper@healthurl.com>
> wrote:
>
>> It’s fundamentally unfair to restrict my use of technology if I’m willing
>> to take full responsibility for the posting. My reputation should suffer
>> just as much if a post offends regardless of what tools I may have used.
>>
>> The problem seems to be that we have no way to enforce human
>> responsibility.
>>
>> As I see it, this is the only problem. I wish we were discussing
>> solutions.
>>
>> Adrian
>>
>> On Tue, Apr 21, 2026 at 9:04 AM Amir Hameed <amsaalegal@gmail.com> wrote:
>>
>>>
>>> We are discussing decentralised standards on a centralised email mailing
>>> list which is open to receive anything , it worked earlier because there
>>> was a limited capability a user had in terms of what they could research,
>>> type an email, structure it well and then send it to the mailing list, we
>>> had very few people who really were willing to put their work and time and
>>> help develop standards , few years back the same user has been handed over
>>> a tool where he can write a sentence and get multiple paragraphs answer
>>> that too structured in a intelligent way but may not be factual, it’s
>>> obvious that users who ever wished to write an email to the mailing list
>>> but could not do that due to lack of both energy to research , draft and
>>> put it forward for discussion might think of using these tools to overcome
>>> that barrier to entry, it’s similar to industry revolution, there was a
>>> time when only elite could afford a car because there was no assembly line
>>> and it was done with hands manually , once we had assembly lines anyone
>>> could buy a car if they had money.
>>>
>>> Our current technology has reached another assembly line moment, this
>>> time it’s not cars but human skills, reasoning ,  and information systems.
>>> So this points us to something deeper and that is we need to rethink the
>>> entire process now, patching doesn’t always help like Kyle said ,
>>> reputation is not helpful in open ecosystems , we may have to elevate the
>>> criteria of what is valuable once intelligence and skills become a
>>> commodity and we need to think of humans as artists in the industrial
>>> world. Technology is not always the only answer , before we decide anything
>>> , let’s step back and rethink how the whole thing has changed ever since
>>> intelligence became commodity and generative tools became digital
>>> replacement of human skills. We may not have the mailing list itself in
>>> future , transition period is always chaotic and we collectively navigate
>>> it, I strongly believe for a better solution we need to rethink and come up
>>> with some fresh perspectives like verifiable provenance, proof of
>>> expertise, proof human , otherwise human signal will drown in this
>>> asymmetry.
>>>
>>>
>>> PS: it’s written by me no tool was used in this except the mail itself ,
>>> It took me few more minutes but it’s worth it
>>>
>>> Regards
>>>
>>>
>>> On Tue, 21 Apr 2026 at 11:27 AM, Kyle Den Hartog <kyle@pryvit.tech>
>>> wrote:
>>>
>>>> Reputation systems work well as a heuristic metric when you’re
>>>> operating in high re-interaction environments. That’s not really the case
>>>> on the Web because of its openness properties where it's easy to build up
>>>> and spend down identities in an automated fashion. It's made even easier
>>>> with LLMs now too.
>>>>
>>>> For example, on this mailing list spammers could form new emails in
>>>> seconds and form new identities to continue their attacks. If you set up a
>>>> guard to prevent it you've now accepted the tradeoff of reduce openness and
>>>> entered a cat and mouse game at the same time. There are discourse forums
>>>> (polkadot and ZCash are 2 examples where I've encountered this) that have
>>>> these techniques built in where you can only post once you’ve built up a
>>>> reputation. They have specific threads that allow people with low
>>>> reputation to engage and then you earn reputation over time. This comes
>>>> with the tradeoff of reducing the openness of the system in exchange for a
>>>> higher bar of entry. Maybe a poster has something legitimate to add to the
>>>> conversation, but because they didn't build their reputation up enough they
>>>> can't contribute. With automation like LLMs given to attackers these days,
>>>> it's producing an asymmetric attack surface and reverting the solution more
>>>> towards option one (Dark Forest theory - retreat to safe communication
>>>> channels).
>>>>
>>>> Another example where we're dealing with these sorts of low value
>>>> sybils is in Brave's hackerone bug bounty programs. There's evidence[1]
>>>> from BugCrowd this could be security vendors using this to gather training
>>>> data, but it also simply could be someone operating out of a lower wage
>>>> country where one bug bounty report can be worth a month's salary or more.
>>>> So they're incentivized to use an LLM to generate new identities on the
>>>> fly, spam bug bounty programs, and if their signal degrades too much they
>>>> drop and swap them.
>>>>
>>>> Additionally, I’m not sure how much you’ve been following the Web3 and
>>>> public goods funding/DAO spaces, but they’ve actually been relying on these
>>>> identity credential systems as a sybil resistance mechanism for a bit now.
>>>> While there’s been mild success shown, the system over time has had to add
>>>> capabilities to address different attacks that have been conducted. For
>>>> example, Gitcoin Grants 24 saw a 60% reduction in sybil attack influence
>>>> from their GG23 round[2]. They’re the most widely deployed system that I’ve
>>>> seen trying to actively go down the route of identity based protections for
>>>> Sybil attacks and spam. Worth a look for you at least but it's also worth
>>>> pointing out they're producing a system that structurally still faces the
>>>> problem as long as the incentives for conducting the attack are still high
>>>> enough ($1.8 million dollars was given out in GG24). For their system they
>>>> rely on over 20 different potential signals including government IDs,
>>>> biometrics, social signals, and financial signals (Binance accounts which
>>>> require KYC)[3]. Even then, people are still successfully conducting
>>>> attacks against this system and as more systems are built on the same
>>>> identity credential based sybil resistances (aka the reputation system atop
>>>> it) the value of conducting a sybil attack grows because it can be
>>>> repurposed across multiple systems.
>>>>
>>>> There's 2 other deployed identity credential systems that have also
>>>> been working on this problem as well in the Web3 space with some issues.
>>>> Idena[4] and Worldcoin[5] have fallen susceptible to some form of Sybil
>>>> attacks also. From what I've seen, people are conducting "puppeteer
>>>> attacks" where one person "puppets" many people who have digital IDs to
>>>> coordinate in the system and conduct attacks. These typically occur
>>>> through an attacker paying for some action to be taken in order to conduct
>>>> the attack. Again, these attacks are usually successful because they're
>>>> operating out of lower wage countries where the seemingly smaller amount of
>>>> money paid makes the attack worth it.
>>>>
>>>> The point here is that attaching reputation systems onto this means
>>>> you're in for a attack surface that has historically struggled to keep up.
>>>> I'm not convinced that an email list is ready to deal with this let alone
>>>> technology built through a standardization process that takes years to
>>>> iterate on. Especially when the human(s) who are participating is actively
>>>> coordinating with agents to conduct the spam or sybil attacks. So yeah,
>>>> that's why I'm not really convinced identity credentials are going to be
>>>> that useful. I'd be happy to be wrong, but what I'm seeing both in terms of
>>>> real world adoption as well as attacks I've had to deal with (we've seen
>>>> these sybil attacks against other systems in Brave too) identity
>>>> credentials only go so far in solving the problem and they come with
>>>> tradeoffs that normally aren't worth it.
>>>>
>>>> Here's some links for the citations made above as well.
>>>> [1] Bugcrowd:
>>>> https://www.bugcrowd.com/blog/bugcrowd-policy-changes-to-address-ai-slop-submissions/
>>>> [2] Gitcoin reduces attacks:
>>>> https://gitcoin.co/research/quadratic-funding-sybil-resistance
>>>> [3] Gitcoin Signals:
>>>> https://support.passport.xyz/passport-knowledge-base/stamps/how-do-i-add-passport-stamps/the-government-id-stamp
>>>> [4] Idena:
>>>> https://stanford-jblp.pubpub.org/pub/compressed-to-0-proof-personhood/release/5
>>>> [5] Worldcoin:
>>>> https://www.dlnews.com/articles/regulation/singapore-officials-warns-against-worldcoin-account-trading/
>>>>
>>>> -Kyle
>>>> -------- Original Message --------
>>>> On Tuesday, 04/21/26 at 05:16 Casanova, Juan <J.Casanova@hw.ac.uk>
>>>> wrote:
>>>>
>>>> Kyle,
>>>>
>>>> You say
>>>>
>>>> Identity credentials are highly unlikely to stop this either which I
>>>> suspect is where many in this community would want to turn. Identity
>>>> credentials just turn the issue back into a key management problem and we
>>>> don’t really have a great way to prevent a user from sharing their keys
>>>> with their agent. That problem persists whether the system has a delegation
>>>> solution or not too.
>>>>
>>>> I think there may be an important "but" to this. I think some of the
>>>> things you suggest later may relate to it, or some of the ideas that Will
>>>> discussed later. I'm definitely sure that there has been much more
>>>> discussion about things like this and more attempted approaches to similar
>>>> things that I am aware, as I still consider myself a newbie here. However,
>>>> let me state my view...
>>>>
>>>> While you can't prevent a user from sharing their keys with their
>>>> agent, you can have, like you said "pseudo-reputation" systems attached to
>>>> keys, that take time and good contributions to build, and are deteriorated
>>>> when providing lower quality contributions. I believe this can be achieved
>>>> without systematically breaking sovereignty. These hypothetical system(s)
>>>> could span across multiple mediums, they don't need to be constrained to
>>>> single contexts, and be optional and complementary rather than strictly
>>>> enforced, but they could help both as a deterrent for people haphazardly
>>>> sharing unfiltered AI contents (I refuse to use the word slop because I
>>>> feel it has connotations that challenge civil conversations and is pretty
>>>> much a slur, even if I understand what people mean by it), and as a way for
>>>> people to identify and neutralize persistent sources of it.
>>>>
>>>> In my view, this is no different to what we already do in our physical
>>>> embodied life. We have face recognition embedded into us (most of us), and
>>>> we learn to create an internal opinion of other people based on their
>>>> interactions with us. When somebody consistently steals our time with
>>>> pointless drivel and unfiltered contributions, we don't need to put them in
>>>> jail, put a sign over their heads that says they are unworthy, or
>>>> (generally speaking) prohibit them from participating in public life. We
>>>> simply don't pay as much attention to them, because we know who they are
>>>> and what their usual approach to contributions is. Identity online simply
>>>> can replace the face recognition in a way that is more flexible and
>>>> preserves sovereignty better, as well as being better equipped to deal with
>>>> the volume.
>>>>
>>>> As I said, I'm sure I am unaware of the extent to which similar ideas
>>>> have been proposed and explored. I am also very aware that in the same way
>>>> that some people here are using questionable predictions of what AI *
>>>> will become *that, whether grounded or not, remain just a prediction
>>>> and not a current reality that can be wielded as a definitive argument for
>>>> what to do right now; what I am discussing here is also a prediction or a
>>>> hope, rather than a current reality. But in the same way that I think it's
>>>> valid to work towards better AI tools, I think it's valid to work towards
>>>> systems that enable us to better * filter through the ocean of
>>>> information* in ways that respect sovereignty for all sides involved,
>>>> can be personalized, and respect our own intelligence. I think it's a dream
>>>> worth pursuing, and I believe it relates directly to the current matter.
>>>>
>>>> But in the meantime, I feel that discussing like we are doing seems to
>>>> already be shaping a lot of moderate people's views into compromises that
>>>> may make this mailing list more comfortable for everybody involved. One way
>>>> or another, we will find out.
>>>>
>>>> *Juan Casanova Jaquete*
>>>>
>>>> Assistant Professor – School of Engineering and Physical Sciences –
>>>> Data Science GA Programme
>>>>
>>>> *j.casanova@hw.ac.uk* <j.casanova@hw.ac.uk> – Earl Mountbatten
>>>> Building 1.31 (Heriot Watt Edinburgh campus)
>>>>
>>>>
>>>>
>>>> Email is an asynchronous communication method. I do not expect and
>>>> others should not expect immediate replies. Reply at your earliest
>>>> convenience and working hours.
>>>>
>>>>
>>>>
>>>> I am affected by Delayed Sleep Phase Disorder. This means that I am an
>>>> extreme night owl. My work day usually begins at 14:00 Edinburgh time, and
>>>> I often work late into the evening and on weekends. Please try to take this
>>>> into account where possible.
>>>>
>>>>
>>>>
>>>>
>>>> ------------------------------
>>>> *From:* Kyle Den Hartog <kyle@pryvit.tech>
>>>> *Sent:* Sunday, April 19, 2026 06:28
>>>> *To:* Steve Capell <steve.capell@gmail.com>
>>>> *Cc:* Melvin Carvalho <melvincarvalho@gmail.com>; Marcus Engvall <
>>>> marcus@engvall.email>; Manu Sporny <msporny@digitalbazaar.com>;
>>>> public-credentials@w3.org <public-credentials@w3.org>
>>>> *Subject:* Re: The Slopification of the CCG
>>>>
>>>> You don't often get email from kyle@pryvit.tech. Learn why this is
>>>> important <https://aka.ms/LearnAboutSenderIdentification>
>>>> ****************************************************************
>>>> Caution: This email originated from a sender outside Heriot-Watt
>>>> University.
>>>> Do not follow links or open attachments if you doubt the authenticity
>>>> of the sender or the content.
>>>> ****************************************************************
>>>>
>>>>
>>>> In case it helps, here’s how things are going in terms of AIPREFs WG
>>>> and the impact on search crawlers:
>>>>
>>>> https://x.com/grittygrease/status/2044152662673752454?s=20
>>>>
>>>> In other words, we don’t really have any enforcement mechanisms here to
>>>> stop this. In fact I highly suspect some people are using them in this
>>>> conversation right now unless their writing styles dramatically changed in
>>>> the past few years. My email client started noticing it via machine
>>>> learning I suspect and filtering threads to my spam inbox like this most of
>>>> the time given I engage a lot less these days. Personally that’s been a
>>>> good enough solution for me.
>>>>
>>>> Identity credentials are highly unlikely to stop this either which I
>>>> suspect is where many in this community would want to turn. Identity
>>>> credentials just turn the issue back into a key management problem and we
>>>> don’t really have a great way to prevent a user from sharing their keys
>>>> with their agent. That problem persists whether the system has a delegation
>>>> solution or not too.
>>>>
>>>> So where do we go? I’m not exactly sure. Here’s the leading theories
>>>> and their tradeoffs that stand out to me for the generalized solution of AI
>>>> generated content:
>>>>
>>>> 1. https://www.ystrickler.com/the-dark-forest-theory-of-the-internet/
>>>> - users just stop engaging in these spaces and retreat to closed door
>>>> forums. Then we lose the open collaboration that made the Web great.
>>>>
>>>> 2. Re-hash DRM debate by making it so users can’t actually access their
>>>> keys used to sign their identity credentials. This seems to be the current
>>>> path governments like. It optimizes enforcement but also entrenches access
>>>> to the Web around a select number of OSes and reduces who’s allowed to
>>>> access and contribute to conversations on the Web. I also see that as a bit
>>>> short sighted.
>>>>
>>>> 3. Re-introduce fingerprinting (and pseudo reputation to that
>>>> fingerprint) based identity like what CAPTCHAs do. That works well for
>>>> service side enforcement but in mailing lists like these not so much. So
>>>> likely will need user controlled filtering like what spam filters in email
>>>> clients do as well.
>>>>
>>>> 4. Is the most interesting but most unproven. We shift how people are
>>>> reachable and build out Horton Protocol like what Mark Miller proposed
>>>> years ago at ActivityPub conf. They may have already tried this and had
>>>> issues. I’m not exactly sure:
>>>> https://www.youtube.com/watch?v=NAfjEnu6R2g
>>>>
>>>> In any case though, we don’t have much of a solution right now in our
>>>> particular forum and outside things like 3, I don’t expect much to change
>>>> in a coordinated manner right now. Looking forward to seeing what we come
>>>> up with though over the next decade and hopefully the trade offs we make
>>>> don’t take away too much of what originally made the Web great.
>>>>
>>>> -Kyle
>>>>
>>>>
>>>> -------- Original Message --------
>>>> On Sunday, 04/19/26 at 13:10 Steve Capell <steve.capell@gmail.com>
>>>> wrote:
>>>>
>>>> Challenge : there’s an increasing amount of AI generated content that,
>>>> whilst possibly containing useful insights, takes more time to read than to
>>>> generate and, given the size of this mailing list, is likely to lead most
>>>> of us to unsubscribe, rendering the list worthless
>>>>
>>>> Constraint : AI used well is a genuinely useful tool and can
>>>> dramatically improve quality of output.  “Used well” is key and,
>>>> unfortunately, many do not use it so well.  Nevertheless, this group can’t
>>>> become anti-LLM luddites or this list may equally become worthless for the
>>>> opposite reason
>>>>
>>>> Goal : to continue to enjoy intelligent discussions between real humans
>>>> that feel empowered to use AI to improve the value of their human
>>>> contributions.  So the goal, it seems to me is not to block AI content but
>>>> rather to block content that has little evidence of human analysis and
>>>> interpretation.  Perhaps counterintuitively, LLMs themselves might be the
>>>> best tool to detect such content
>>>>
>>>> Proposal : rather than continuing to discuss whether AI content on this
>>>> list is good or bad, let’s collectively agree a rubric in the form of an AI
>>>> prompt that can act as an automated list moderator.  The rubric should
>>>> focus on requiring evidence of human assessment rather than blocking AI
>>>> content
>>>>
>>>> I had a go at this myself with several of the messages in this thread
>>>> and earlier ones and it seemed quite effective at blocking the ones that I
>>>> would have blocked myself.  I know that there is a token cost associated
>>>> with such a moderator but I for one would delighted to contribute.
>>>>
>>>> Disclaimer : this message was written with blurry eyes and fat thumbs
>>>> on my iPhone - with no AI assistance whatsoever
>>>>
>>>> Kind regards
>>>>
>>>> Steven Capell
>>>> UN/CEFACT Vice-Chair
>>>> Mob: +61 410 437854
>>>>
>>>> On 19 Apr 2026, at 10:03 am, Melvin Carvalho <melvincarvalho@gmail.com>
>>>> wrote:
>>>>
>>>> 
>>>>
>>>>
>>>> ne 19. 4. 2026 v 1:49 odesílatel Marcus Engvall <marcus@engvall.email>
>>>> napsal:
>>>>
>>>> Hi all,
>>>>
>>>> I’m glad to see that we have some healthy discourse in this thread with
>>>> a variety of views. I would like to address some of the points made.
>>>>
>>>> On 18 Apr 2026, at 01:50, Melvin Carvalho <melvincarvalho@gmail.com>
>>>> wrote:
>>>>
>>>> LLMs have the advantage that they know most or all of the specs
>>>> inside-out, due to their training. Most humans (with notable exceptions),
>>>> including on this list, have partial understanding of the complete works of
>>>> web standards.
>>>>
>>>>
>>>> This is a real advantage that these tools have and it should not be
>>>> understated. I use them professionally for referential lookups and for
>>>> confirming hypotheses, and I have no doubt that they have the ability to
>>>> accelerate otherwise excellent standards work. But I am also careful to not
>>>> fall into the trap of assuming that their lexical consistency can fully
>>>> substitute  for human judgement. LLMs are probabilistic models with
>>>> encyclopaedic knowledge, they are not deterministic oracles with the
>>>> capacity to rigorously derive that same knowledge. In the context of the
>>>> kind of work done in this group I think it is important to not confuse the
>>>> two. I trust an LLM to give me a comprehensive overview of a standards
>>>> framework - I do not, however, trust it to prescribe the framework itself
>>>> without and human review and editorial judgement.
>>>>
>>>> I do however concede on your point on testing methodology, and I think
>>>> you raise a good point that Manu eloquently touched on.
>>>>
>>>>
>>>> Good points. However LLMs outperform humans on medical exams,
>>>> olympiad questions and many other tests, often by wide margins. They are
>>>> much more than prediction machines or probabilistic guessers. What I'm
>>>> saying is that I predict LLMs would exceed humans in the standards setting
>>>> on any quantitative evaluation. We just have not the tools to evaluate yet.
>>>> However, I believe the picture will be much clearer one year from now.
>>>>
>>>>
>>>>
>>>> On 18 Apr 2026, at 02:24, Manu Sporny <msporny@digitalbazaar.com>
>>>> wrote:
>>>>
>>>> Technology transitions, especially ones around human communication can
>>>> be rough to navigate. This one is no different, and sometimes it takes
>>>> decades to figure out the norms around a new medium (the printed page,
>>>> radio, television, BBSes, mailing lists, AOL, ICQ, Napster, Twitter,
>>>> Digg/Reddit/Discord, and so on).
>>>>
>>>>
>>>> You are completely right that this is a transition, and I think we are
>>>> all trying to map this new technology onto our existing mental models of
>>>> what discourse should and could be. Friction and contention is bound to
>>>> arise. It is clearly counterproductive, as you and later Amir rightly
>>>> stated, to enforce neo-Luddism and reject the technology wholesale.
>>>>
>>>> My point however is that the ability to passively follow and
>>>> occasionally contribute to developments and discussions in this group is
>>>> immensely valuable, both commercially and technically. Compressing the
>>>> signal-to-noise ratio raises the bar for both comprehension and
>>>> participation, and my fear is that the inevitable intractability will, as
>>>> you pointed out in the other thread, overwhelm people and alienate them,
>>>> especially those of us with many other commitments and who do not have the
>>>> time or ability to participate in every group call. That said, it is,
>>>> as you suggested, our responsibility to moderate our own information
>>>> ingestion, as has been the case for time immemorial in any rhetorical forum.
>>>>
>>>> Perhaps LLMs will simply change the structure of how discourse is
>>>> conducted in forums like these rather than drown it out, as some other
>>>> writers have suggested in the thread. If the cost to contribute text tends
>>>> to zero, naturally the valuable discussions will shift elsewhere to forums
>>>> that still have a cost, such as the group calls. I just hope the work
>>>> doesn’t lose the diversity of opinions that is crucial to develop a refined
>>>> and well-considered standard.
>>>>
>>>> --
>>>> Marcus Engvall
>>>>
>>>> Principal—M. Engvall & Co.
>>>> mengvall.com
>>>>
>>>> ------------------------------
>>>>
>>>> Founded in 1821, Heriot-Watt is a leader in ideas and solutions. With
>>>> campuses and students across the entire globe we span the world, delivering
>>>> innovation and educational excellence in business, engineering, design and
>>>> the physical, social and life sciences. This email is generated from the
>>>> Heriot-Watt University Group, which includes:
>>>>
>>>>    1. Heriot-Watt University, a Scottish charity registered under
>>>>    number SC000278
>>>>    2. Heriot- Watt Services Limited (Oriam), Scotland's national
>>>>    performance centre for sport. Heriot-Watt Services Limited is a private
>>>>    limited company registered is Scotland with registered number SC271030 and
>>>>    registered office at Research & Enterprise Services Heriot-Watt University,
>>>>    Riccarton, Edinburgh, EH14 4AS.
>>>>
>>>> The contents (including any attachments) are confidential. If you are
>>>> not the intended recipient of this e-mail, any disclosure, copying,
>>>> distribution or use of its contents is strictly prohibited, and you should
>>>> please notify the sender immediately and then delete it (including any
>>>> attachments) from your system.
>>>>
>>>>

Received on Thursday, 23 April 2026 17:27:00 UTC