- From: Bob Wyman <bob@wyman.us>
- Date: Thu, 23 Apr 2026 14:30:25 -0400
- To: Alan Karp <alanhkarp@gmail.com>
- Cc: Moses Ma <moses.ma@futurelabconsulting.com>, NIKOLAOS FOTIOY <fotiou@aueb.gr>, Amir Hameed <amsaalegal@gmail.com>, Adrian Gropper <agropper@healthurl.com>, Kyle Den Hartog <kyle@pryvit.tech>, Juan Casanova <j.casanova@hw.ac.uk>, Steve Capell <steve.capell@gmail.com>, Melvin Carvalho <melvincarvalho@gmail.com>, Marcus Engvall <marcus@engvall.email>, Manu Sporny <msporny@digitalbazaar.com>, Public-Credentials <public-credentials@w3.org>
- Message-ID: <CAA1s49W3PuEqnP5ns34usOBgj8Jbt22DLuKQXXjMAUJ9DLL4Lg@mail.gmail.com>
Alan wrote: "in the interest of people with less time to read than I, any post longer than a paragraph or two should start with a tldr." Unfortunately, the utility of concision is inversely proportional to the novelty of the expression. But, while that sentence is fairly concise, if I didn't know the word "concision," or if I assumed you didn't, I might have written many more words, e.g. "The utility of expressing ideas clearly, using the fewest necessary words,..." bob wyman On Thu, Apr 23, 2026 at 1:29 PM Alan Karp <alanhkarp@gmail.com> wrote: > Not to denigrate anyone's posts, including this one, but one complaint > seems to be the length of posts contributed by AI. I find it ironic that > many such posts are quite long. I personally find that content > interesting, but perhaps, in the interest of people with less time to read > than I, any post longer than a paragraph or two should start with a tldr. > > -------------- > Alan Karp > > > On Thu, Apr 23, 2026 at 10:20 AM Moses Ma < > moses.ma@futurelabconsulting.com> wrote: > >> Nikos, >> >> I wanted to applaud your candor, which was refreshingly straightforward. >> However, I urge everyone to adhere to the unspoken goal of inclusivity >> here. We need to be peacemakers. >> >> Anyway, your post primed the pump of ideas. >> >> 1) A small experiment I ran recently showed ~25% of people in this forum >> can’t reliably distinguish LLM output from human writing. So some of what >> gets labeled “AI slop” is actually just perception. But that’s almost >> beside the point here. The real issue isn’t AI—it’s verbosity as a >> strategy. >> >> 2) Standards groups generally follow a few structural dynamics: >> >> - inclusion > exclusion >> - visibility = influence >> - language = control >> >> What’s changed is that AI has reduced the cost of writing. So individuals >> who were already inclined to, ah, “over-contribute” can now scale that >> behavior, flooding the channel with low-signal, self-promotional, or >> tangential content. These people are like the blowhards at a company who >> believe that talking really loud and constantly mansplaining is a success >> strategy. This is not the kind of leadership we need in the 21st century. >> >> The failure mode isn’t just annoyance—it’s attention capture by volume, >> where a few verbose participants degrade signal-to-noise for everyone else. >> What we really need is an AI with a *bloviation sensor*. Most of us do >> this internally by simply not reading certain posts, until we’ve had enough >> and lash out. Then the bloviator is justified in feeling attacked. >> >> Therefore, instead of debating tools or reputation, it may be more >> productive to consider the possibility of placing lightweight guardrails on >> contribution quality: >> >> - Contribution caps per cycle (forces prioritization) >> - One idea per message (no multi-topic dumps or longwinded responses >> like this one) >> - Editorial compression rights (chairs or AI can edit and summarize >> without loss of weight) >> - Track signal-to-noise to reward reputation for increased caps (what >> actually survives into the draft) >> >> Something like this would get us out of policing individuals or tools, >> and back to protecting the quality of the work. >> >> 3) I think that in a few years, people who refuse to use LLMs—essentially >> “artisanal writers”—will seem as quaint as luddites who refuse to use >> Google. I added the em dashes to show that a human generated response can >> still use em dashes, they are not the six fingers of LLM writing. I like >> them because they force the mind to “slow the breath” while reading. >> >> To wrap up, the happy ending I’d love to see is something likely >> impossible. My preference is a new kind of process that nurtures growth, by >> encouraging the hesitant to find their voice, novices to get up to speed >> faster, and the emergence of greater self-awareness by the bloviators. >> >> I’m actually working on a Web 89.0 vision of this vision (haha)…. >> >> My incubator is building a “stealth-ish mode” startup called EmergentYOU >> with the goal of creating something that could provide a labor transition >> cushion for the AI era—using longitudinal coaching, hyperpersonalized >> learning pathways, and an AI career co-pilot to continuously align people >> with opportunity. It converts disruption into mobility by linking skills, >> employers, and outcomes in a closed-loop system that compounds human >> potential over time. We’ll announce the EmergentYOU concept at Human >> Tech Week in San Francisco next month. >> >> Anyway, I’ve been thinking about we’re extending EmergentYOU into >> EmergentUS, to support teams: an intelligence layer designed to increase group >> cohesion, reduce participation disparity, and enhance group consonance. >> The system nudges quieter participants to contribute, modulates dominant >> voices, and elevates high-signal input—creating balanced, adaptive dialogue >> and measurably stronger collective performance. However, the real issue is >> that the entire team would need to agree to undergo the process. If there >> is interest in piloting EmergentUS in an SDO context… let me know. >> >> – Moses >> >> >> PS, if you’re in the SF Bay Area and would like to attend our event at >> Human Tech Week… let me know too. >> >> >> <moses@nureon-eda.ai> >> >> On Apr 23, 2026 at 12:55 AM, <NIKOLAOS FOTIOY <fotiou@aueb.gr>> wrote: >> >> Hi all, >> I think we are losing the context here. The problem is that certain >> perfectly identifiable individuals spam the list with mostly meaningless, >> self-promotional content. For example, every couple of messages I receive >> some web 7.0 irrelevant, non sense. AI tools have just made their job >> easier to generate content. Blaming AI tools is just a polite way to say to >> those individuals “please stop you are creating too much noise” >> >> Best, >> Nikos >> >> 23 Απρ 2026, 10:36 πμ, ο χρήστης «Amir Hameed <amsaalegal@gmail.com>» >> έγραψε: >> >> >> Hi Adrian, >> >> I do not think the concern is about restricting the use of tools. People >> will use whatever tools are available to them—that’s inevitable. >> >> The issue is that reputation alone is not a strong enough primitive for >> systems that aim to operate at scale and across jurisdictions. >> >> In distributed environments, we typically rely on properties like: >> >> - verifiable provenance >> - non-repudiation >> - integrity of authorship >> >> These are not about limiting expression, but about ensuring that >> contributions can be evaluated independent of the individual’s perceived >> credibility. >> >> Saying “my reputation will suffer if I’m wrong” assumes: >> >> 1. reputations are consistently observable across contexts, and >> 2. reputational consequences are sufficient deterrent >> >> In practice, neither assumption holds reliably—especially in global, >> asynchronous systems. >> >> On enforcement: Global enforcement is not realistic. That’s precisely why >> systems tend to push guarantees down to verifiable layers rather than >> relying on behavioral expectations at the application layer. >> >> So perhaps the problem is not tool usage vs. responsibility, but: >> >> how do we make authorship and intent more verifiable without constraining >> participation? >> >> Regards >> Amir Hameed >> >> On Thu, 23 Apr 2026 at 12:15 PM, Adrian Gropper <agropper@healthurl.com> >> wrote: >> >>> It’s fundamentally unfair to restrict my use of technology if I’m >>> willing to take full responsibility for the posting. My reputation should >>> suffer just as much if a post offends regardless of what tools I may have >>> used. >>> >>> The problem seems to be that we have no way to enforce human >>> responsibility. >>> >>> As I see it, this is the only problem. I wish we were discussing >>> solutions. >>> >>> Adrian >>> >>> On Tue, Apr 21, 2026 at 9:04 AM Amir Hameed <amsaalegal@gmail.com> >>> wrote: >>> >>>> >>>> We are discussing decentralised standards on a centralised email >>>> mailing list which is open to receive anything , it worked earlier because >>>> there was a limited capability a user had in terms of what they could >>>> research, type an email, structure it well and then send it to the mailing >>>> list, we had very few people who really were willing to put their work and >>>> time and help develop standards , few years back the same user has been >>>> handed over a tool where he can write a sentence and get multiple >>>> paragraphs answer that too structured in a intelligent way but may not be >>>> factual, it’s obvious that users who ever wished to write an email to the >>>> mailing list but could not do that due to lack of both energy to research , >>>> draft and put it forward for discussion might think of using these tools to >>>> overcome that barrier to entry, it’s similar to industry revolution, there >>>> was a time when only elite could afford a car because there was no assembly >>>> line and it was done with hands manually , once we had assembly lines >>>> anyone could buy a car if they had money. >>>> >>>> Our current technology has reached another assembly line moment, this >>>> time it’s not cars but human skills, reasoning , and information systems. >>>> So this points us to something deeper and that is we need to rethink the >>>> entire process now, patching doesn’t always help like Kyle said , >>>> reputation is not helpful in open ecosystems , we may have to elevate the >>>> criteria of what is valuable once intelligence and skills become a >>>> commodity and we need to think of humans as artists in the industrial >>>> world. Technology is not always the only answer , before we decide anything >>>> , let’s step back and rethink how the whole thing has changed ever since >>>> intelligence became commodity and generative tools became digital >>>> replacement of human skills. We may not have the mailing list itself in >>>> future , transition period is always chaotic and we collectively navigate >>>> it, I strongly believe for a better solution we need to rethink and come up >>>> with some fresh perspectives like verifiable provenance, proof of >>>> expertise, proof human , otherwise human signal will drown in this >>>> asymmetry. >>>> >>>> >>>> PS: it’s written by me no tool was used in this except the mail itself >>>> , It took me few more minutes but it’s worth it >>>> >>>> Regards >>>> >>>> >>>> On Tue, 21 Apr 2026 at 11:27 AM, Kyle Den Hartog <kyle@pryvit.tech> >>>> wrote: >>>> >>>>> Reputation systems work well as a heuristic metric when you’re >>>>> operating in high re-interaction environments. That’s not really the case >>>>> on the Web because of its openness properties where it's easy to build up >>>>> and spend down identities in an automated fashion. It's made even easier >>>>> with LLMs now too. >>>>> >>>>> For example, on this mailing list spammers could form new emails in >>>>> seconds and form new identities to continue their attacks. If you set up a >>>>> guard to prevent it you've now accepted the tradeoff of reduce openness and >>>>> entered a cat and mouse game at the same time. There are discourse forums >>>>> (polkadot and ZCash are 2 examples where I've encountered this) that have >>>>> these techniques built in where you can only post once you’ve built up a >>>>> reputation. They have specific threads that allow people with low >>>>> reputation to engage and then you earn reputation over time. This comes >>>>> with the tradeoff of reducing the openness of the system in exchange for a >>>>> higher bar of entry. Maybe a poster has something legitimate to add to the >>>>> conversation, but because they didn't build their reputation up enough they >>>>> can't contribute. With automation like LLMs given to attackers these days, >>>>> it's producing an asymmetric attack surface and reverting the solution more >>>>> towards option one (Dark Forest theory - retreat to safe communication >>>>> channels). >>>>> >>>>> Another example where we're dealing with these sorts of low value >>>>> sybils is in Brave's hackerone bug bounty programs. There's evidence[1] >>>>> from BugCrowd this could be security vendors using this to gather training >>>>> data, but it also simply could be someone operating out of a lower wage >>>>> country where one bug bounty report can be worth a month's salary or more. >>>>> So they're incentivized to use an LLM to generate new identities on the >>>>> fly, spam bug bounty programs, and if their signal degrades too much they >>>>> drop and swap them. >>>>> >>>>> Additionally, I’m not sure how much you’ve been following the Web3 and >>>>> public goods funding/DAO spaces, but they’ve actually been relying on these >>>>> identity credential systems as a sybil resistance mechanism for a bit now. >>>>> While there’s been mild success shown, the system over time has had to add >>>>> capabilities to address different attacks that have been conducted. For >>>>> example, Gitcoin Grants 24 saw a 60% reduction in sybil attack influence >>>>> from their GG23 round[2]. They’re the most widely deployed system that I’ve >>>>> seen trying to actively go down the route of identity based protections for >>>>> Sybil attacks and spam. Worth a look for you at least but it's also worth >>>>> pointing out they're producing a system that structurally still faces the >>>>> problem as long as the incentives for conducting the attack are still high >>>>> enough ($1.8 million dollars was given out in GG24). For their system they >>>>> rely on over 20 different potential signals including government IDs, >>>>> biometrics, social signals, and financial signals (Binance accounts which >>>>> require KYC)[3]. Even then, people are still successfully conducting >>>>> attacks against this system and as more systems are built on the same >>>>> identity credential based sybil resistances (aka the reputation system atop >>>>> it) the value of conducting a sybil attack grows because it can be >>>>> repurposed across multiple systems. >>>>> >>>>> There's 2 other deployed identity credential systems that have also >>>>> been working on this problem as well in the Web3 space with some issues. >>>>> Idena[4] and Worldcoin[5] have fallen susceptible to some form of Sybil >>>>> attacks also. From what I've seen, people are conducting "puppeteer >>>>> attacks" where one person "puppets" many people who have digital IDs to >>>>> coordinate in the system and conduct attacks. These typically occur >>>>> through an attacker paying for some action to be taken in order to conduct >>>>> the attack. Again, these attacks are usually successful because they're >>>>> operating out of lower wage countries where the seemingly smaller amount of >>>>> money paid makes the attack worth it. >>>>> >>>>> The point here is that attaching reputation systems onto this means >>>>> you're in for a attack surface that has historically struggled to keep up. >>>>> I'm not convinced that an email list is ready to deal with this let alone >>>>> technology built through a standardization process that takes years to >>>>> iterate on. Especially when the human(s) who are participating is actively >>>>> coordinating with agents to conduct the spam or sybil attacks. So yeah, >>>>> that's why I'm not really convinced identity credentials are going to be >>>>> that useful. I'd be happy to be wrong, but what I'm seeing both in terms of >>>>> real world adoption as well as attacks I've had to deal with (we've seen >>>>> these sybil attacks against other systems in Brave too) identity >>>>> credentials only go so far in solving the problem and they come with >>>>> tradeoffs that normally aren't worth it. >>>>> >>>>> Here's some links for the citations made above as well. >>>>> [1] Bugcrowd: >>>>> https://www.bugcrowd.com/blog/bugcrowd-policy-changes-to-address-ai-slop-submissions/ >>>>> [2] Gitcoin reduces attacks: >>>>> https://gitcoin.co/research/quadratic-funding-sybil-resistance >>>>> [3] Gitcoin Signals: >>>>> https://support.passport.xyz/passport-knowledge-base/stamps/how-do-i-add-passport-stamps/the-government-id-stamp >>>>> [4] Idena: >>>>> https://stanford-jblp.pubpub.org/pub/compressed-to-0-proof-personhood/release/5 >>>>> [5] Worldcoin: >>>>> https://www.dlnews.com/articles/regulation/singapore-officials-warns-against-worldcoin-account-trading/ >>>>> >>>>> -Kyle >>>>> -------- Original Message -------- >>>>> On Tuesday, 04/21/26 at 05:16 Casanova, Juan <J.Casanova@hw.ac.uk> >>>>> wrote: >>>>> >>>>> Kyle, >>>>> >>>>> You say >>>>> >>>>> Identity credentials are highly unlikely to stop this either which I >>>>> suspect is where many in this community would want to turn. Identity >>>>> credentials just turn the issue back into a key management problem and we >>>>> don’t really have a great way to prevent a user from sharing their keys >>>>> with their agent. That problem persists whether the system has a delegation >>>>> solution or not too. >>>>> >>>>> I think there may be an important "but" to this. I think some of the >>>>> things you suggest later may relate to it, or some of the ideas that Will >>>>> discussed later. I'm definitely sure that there has been much more >>>>> discussion about things like this and more attempted approaches to similar >>>>> things that I am aware, as I still consider myself a newbie here. However, >>>>> let me state my view... >>>>> >>>>> While you can't prevent a user from sharing their keys with their >>>>> agent, you can have, like you said "pseudo-reputation" systems attached to >>>>> keys, that take time and good contributions to build, and are deteriorated >>>>> when providing lower quality contributions. I believe this can be achieved >>>>> without systematically breaking sovereignty. These hypothetical system(s) >>>>> could span across multiple mediums, they don't need to be constrained to >>>>> single contexts, and be optional and complementary rather than strictly >>>>> enforced, but they could help both as a deterrent for people haphazardly >>>>> sharing unfiltered AI contents (I refuse to use the word slop because I >>>>> feel it has connotations that challenge civil conversations and is pretty >>>>> much a slur, even if I understand what people mean by it), and as a way for >>>>> people to identify and neutralize persistent sources of it. >>>>> >>>>> In my view, this is no different to what we already do in our physical >>>>> embodied life. We have face recognition embedded into us (most of us), and >>>>> we learn to create an internal opinion of other people based on their >>>>> interactions with us. When somebody consistently steals our time with >>>>> pointless drivel and unfiltered contributions, we don't need to put them in >>>>> jail, put a sign over their heads that says they are unworthy, or >>>>> (generally speaking) prohibit them from participating in public life. We >>>>> simply don't pay as much attention to them, because we know who they are >>>>> and what their usual approach to contributions is. Identity online simply >>>>> can replace the face recognition in a way that is more flexible and >>>>> preserves sovereignty better, as well as being better equipped to deal with >>>>> the volume. >>>>> >>>>> As I said, I'm sure I am unaware of the extent to which similar ideas >>>>> have been proposed and explored. I am also very aware that in the same way >>>>> that some people here are using questionable predictions of what AI * >>>>> will become *that, whether grounded or not, remain just a prediction >>>>> and not a current reality that can be wielded as a definitive argument for >>>>> what to do right now; what I am discussing here is also a prediction or a >>>>> hope, rather than a current reality. But in the same way that I think it's >>>>> valid to work towards better AI tools, I think it's valid to work towards >>>>> systems that enable us to better * filter through the ocean of >>>>> information* in ways that respect sovereignty for all sides involved, >>>>> can be personalized, and respect our own intelligence. I think it's a dream >>>>> worth pursuing, and I believe it relates directly to the current matter. >>>>> >>>>> But in the meantime, I feel that discussing like we are doing seems to >>>>> already be shaping a lot of moderate people's views into compromises that >>>>> may make this mailing list more comfortable for everybody involved. One way >>>>> or another, we will find out. >>>>> >>>>> *Juan Casanova Jaquete* >>>>> >>>>> Assistant Professor – School of Engineering and Physical Sciences – >>>>> Data Science GA Programme >>>>> >>>>> *j.casanova@hw.ac.uk* <j.casanova@hw.ac.uk> – Earl Mountbatten >>>>> Building 1.31 (Heriot Watt Edinburgh campus) >>>>> >>>>> >>>>> >>>>> Email is an asynchronous communication method. I do not expect and >>>>> others should not expect immediate replies. Reply at your earliest >>>>> convenience and working hours. >>>>> >>>>> >>>>> >>>>> I am affected by Delayed Sleep Phase Disorder. This means that I am an >>>>> extreme night owl. My work day usually begins at 14:00 Edinburgh time, and >>>>> I often work late into the evening and on weekends. Please try to take this >>>>> into account where possible. >>>>> >>>>> >>>>> >>>>> >>>>> ------------------------------ >>>>> *From:* Kyle Den Hartog <kyle@pryvit.tech> >>>>> *Sent:* Sunday, April 19, 2026 06:28 >>>>> *To:* Steve Capell <steve.capell@gmail.com> >>>>> *Cc:* Melvin Carvalho <melvincarvalho@gmail.com>; Marcus Engvall < >>>>> marcus@engvall.email>; Manu Sporny <msporny@digitalbazaar.com>; >>>>> public-credentials@w3.org <public-credentials@w3.org> >>>>> *Subject:* Re: The Slopification of the CCG >>>>> >>>>> You don't often get email from kyle@pryvit.tech. Learn why this is >>>>> important <https://aka.ms/LearnAboutSenderIdentification> >>>>> **************************************************************** >>>>> Caution: This email originated from a sender outside Heriot-Watt >>>>> University. >>>>> Do not follow links or open attachments if you doubt the authenticity >>>>> of the sender or the content. >>>>> **************************************************************** >>>>> >>>>> >>>>> In case it helps, here’s how things are going in terms of AIPREFs WG >>>>> and the impact on search crawlers: >>>>> >>>>> https://x.com/grittygrease/status/2044152662673752454?s=20 >>>>> >>>>> In other words, we don’t really have any enforcement mechanisms here >>>>> to stop this. In fact I highly suspect some people are using them in this >>>>> conversation right now unless their writing styles dramatically changed in >>>>> the past few years. My email client started noticing it via machine >>>>> learning I suspect and filtering threads to my spam inbox like this most of >>>>> the time given I engage a lot less these days. Personally that’s been a >>>>> good enough solution for me. >>>>> >>>>> Identity credentials are highly unlikely to stop this either which I >>>>> suspect is where many in this community would want to turn. Identity >>>>> credentials just turn the issue back into a key management problem and we >>>>> don’t really have a great way to prevent a user from sharing their keys >>>>> with their agent. That problem persists whether the system has a delegation >>>>> solution or not too. >>>>> >>>>> So where do we go? I’m not exactly sure. Here’s the leading theories >>>>> and their tradeoffs that stand out to me for the generalized solution of AI >>>>> generated content: >>>>> >>>>> 1. https://www.ystrickler.com/the-dark-forest-theory-of-the-internet/ >>>>> - users just stop engaging in these spaces and retreat to closed door >>>>> forums. Then we lose the open collaboration that made the Web great. >>>>> >>>>> 2. Re-hash DRM debate by making it so users can’t actually access >>>>> their keys used to sign their identity credentials. This seems to be the >>>>> current path governments like. It optimizes enforcement but also entrenches >>>>> access to the Web around a select number of OSes and reduces who’s allowed >>>>> to access and contribute to conversations on the Web. I also see that as a >>>>> bit short sighted. >>>>> >>>>> 3. Re-introduce fingerprinting (and pseudo reputation to that >>>>> fingerprint) based identity like what CAPTCHAs do. That works well for >>>>> service side enforcement but in mailing lists like these not so much. So >>>>> likely will need user controlled filtering like what spam filters in email >>>>> clients do as well. >>>>> >>>>> 4. Is the most interesting but most unproven. We shift how people are >>>>> reachable and build out Horton Protocol like what Mark Miller proposed >>>>> years ago at ActivityPub conf. They may have already tried this and had >>>>> issues. I’m not exactly sure: >>>>> https://www.youtube.com/watch?v=NAfjEnu6R2g >>>>> >>>>> In any case though, we don’t have much of a solution right now in our >>>>> particular forum and outside things like 3, I don’t expect much to change >>>>> in a coordinated manner right now. Looking forward to seeing what we come >>>>> up with though over the next decade and hopefully the trade offs we make >>>>> don’t take away too much of what originally made the Web great. >>>>> >>>>> -Kyle >>>>> >>>>> >>>>> -------- Original Message -------- >>>>> On Sunday, 04/19/26 at 13:10 Steve Capell <steve.capell@gmail.com> >>>>> wrote: >>>>> >>>>> Challenge : there’s an increasing amount of AI generated content that, >>>>> whilst possibly containing useful insights, takes more time to read than to >>>>> generate and, given the size of this mailing list, is likely to lead most >>>>> of us to unsubscribe, rendering the list worthless >>>>> >>>>> Constraint : AI used well is a genuinely useful tool and can >>>>> dramatically improve quality of output. “Used well” is key and, >>>>> unfortunately, many do not use it so well. Nevertheless, this group can’t >>>>> become anti-LLM luddites or this list may equally become worthless for the >>>>> opposite reason >>>>> >>>>> Goal : to continue to enjoy intelligent discussions between real >>>>> humans that feel empowered to use AI to improve the value of their human >>>>> contributions. So the goal, it seems to me is not to block AI content but >>>>> rather to block content that has little evidence of human analysis and >>>>> interpretation. Perhaps counterintuitively, LLMs themselves might be the >>>>> best tool to detect such content >>>>> >>>>> Proposal : rather than continuing to discuss whether AI content on >>>>> this list is good or bad, let’s collectively agree a rubric in the form of >>>>> an AI prompt that can act as an automated list moderator. The rubric >>>>> should focus on requiring evidence of human assessment rather than blocking >>>>> AI content >>>>> >>>>> I had a go at this myself with several of the messages in this thread >>>>> and earlier ones and it seemed quite effective at blocking the ones that I >>>>> would have blocked myself. I know that there is a token cost associated >>>>> with such a moderator but I for one would delighted to contribute. >>>>> >>>>> Disclaimer : this message was written with blurry eyes and fat thumbs >>>>> on my iPhone - with no AI assistance whatsoever >>>>> >>>>> Kind regards >>>>> >>>>> Steven Capell >>>>> UN/CEFACT Vice-Chair >>>>> Mob: +61 410 437854 >>>>> >>>>> On 19 Apr 2026, at 10:03 am, Melvin Carvalho <melvincarvalho@gmail.com> >>>>> wrote: >>>>> >>>>> >>>>> >>>>> >>>>> ne 19. 4. 2026 v 1:49 odesílatel Marcus Engvall <marcus@engvall.email> >>>>> napsal: >>>>> >>>>> Hi all, >>>>> >>>>> I’m glad to see that we have some healthy discourse in this thread >>>>> with a variety of views. I would like to address some of the points made. >>>>> >>>>> On 18 Apr 2026, at 01:50, Melvin Carvalho <melvincarvalho@gmail.com> >>>>> wrote: >>>>> >>>>> LLMs have the advantage that they know most or all of the specs >>>>> inside-out, due to their training. Most humans (with notable exceptions), >>>>> including on this list, have partial understanding of the complete works of >>>>> web standards. >>>>> >>>>> >>>>> This is a real advantage that these tools have and it should not be >>>>> understated. I use them professionally for referential lookups and for >>>>> confirming hypotheses, and I have no doubt that they have the ability to >>>>> accelerate otherwise excellent standards work. But I am also careful to not >>>>> fall into the trap of assuming that their lexical consistency can fully >>>>> substitute for human judgement. LLMs are probabilistic models with >>>>> encyclopaedic knowledge, they are not deterministic oracles with the >>>>> capacity to rigorously derive that same knowledge. In the context of the >>>>> kind of work done in this group I think it is important to not confuse the >>>>> two. I trust an LLM to give me a comprehensive overview of a standards >>>>> framework - I do not, however, trust it to prescribe the framework itself >>>>> without and human review and editorial judgement. >>>>> >>>>> I do however concede on your point on testing methodology, and I think >>>>> you raise a good point that Manu eloquently touched on. >>>>> >>>>> >>>>> Good points. However LLMs outperform humans on medical exams, >>>>> olympiad questions and many other tests, often by wide margins. They are >>>>> much more than prediction machines or probabilistic guessers. What I'm >>>>> saying is that I predict LLMs would exceed humans in the standards setting >>>>> on any quantitative evaluation. We just have not the tools to evaluate yet. >>>>> However, I believe the picture will be much clearer one year from now. >>>>> >>>>> >>>>> >>>>> On 18 Apr 2026, at 02:24, Manu Sporny <msporny@digitalbazaar.com> >>>>> wrote: >>>>> >>>>> Technology transitions, especially ones around human communication can >>>>> be rough to navigate. This one is no different, and sometimes it takes >>>>> decades to figure out the norms around a new medium (the printed page, >>>>> radio, television, BBSes, mailing lists, AOL, ICQ, Napster, Twitter, >>>>> Digg/Reddit/Discord, and so on). >>>>> >>>>> >>>>> You are completely right that this is a transition, and I think we are >>>>> all trying to map this new technology onto our existing mental models of >>>>> what discourse should and could be. Friction and contention is bound to >>>>> arise. It is clearly counterproductive, as you and later Amir rightly >>>>> stated, to enforce neo-Luddism and reject the technology wholesale. >>>>> >>>>> My point however is that the ability to passively follow and >>>>> occasionally contribute to developments and discussions in this group is >>>>> immensely valuable, both commercially and technically. Compressing the >>>>> signal-to-noise ratio raises the bar for both comprehension and >>>>> participation, and my fear is that the inevitable intractability will, as >>>>> you pointed out in the other thread, overwhelm people and alienate them, >>>>> especially those of us with many other commitments and who do not have the >>>>> time or ability to participate in every group call. That said, it is, >>>>> as you suggested, our responsibility to moderate our own information >>>>> ingestion, as has been the case for time immemorial in any rhetorical forum. >>>>> >>>>> Perhaps LLMs will simply change the structure of how discourse is >>>>> conducted in forums like these rather than drown it out, as some other >>>>> writers have suggested in the thread. If the cost to contribute text tends >>>>> to zero, naturally the valuable discussions will shift elsewhere to forums >>>>> that still have a cost, such as the group calls. I just hope the work >>>>> doesn’t lose the diversity of opinions that is crucial to develop a refined >>>>> and well-considered standard. >>>>> >>>>> -- >>>>> Marcus Engvall >>>>> >>>>> Principal—M. Engvall & Co. >>>>> mengvall.com >>>>> >>>>> ------------------------------ >>>>> >>>>> Founded in 1821, Heriot-Watt is a leader in ideas and solutions. With >>>>> campuses and students across the entire globe we span the world, delivering >>>>> innovation and educational excellence in business, engineering, design and >>>>> the physical, social and life sciences. This email is generated from the >>>>> Heriot-Watt University Group, which includes: >>>>> >>>>> 1. Heriot-Watt University, a Scottish charity registered under >>>>> number SC000278 >>>>> 2. Heriot- Watt Services Limited (Oriam), Scotland's national >>>>> performance centre for sport. Heriot-Watt Services Limited is a private >>>>> limited company registered is Scotland with registered number SC271030 and >>>>> registered office at Research & Enterprise Services Heriot-Watt University, >>>>> Riccarton, Edinburgh, EH14 4AS. >>>>> >>>>> The contents (including any attachments) are confidential. If you are >>>>> not the intended recipient of this e-mail, any disclosure, copying, >>>>> distribution or use of its contents is strictly prohibited, and you should >>>>> please notify the sender immediately and then delete it (including any >>>>> attachments) from your system. >>>>> >>>>>
Received on Thursday, 23 April 2026 18:30:49 UTC