- From: David Booth <david@dbooth.org>
- Date: Thu, 30 Mar 2023 16:14:37 -0400
- To: Adeel <aahmad1811@gmail.com>
- Cc: adasal <adam.saltiel@gmail.com>, Martin Hepp <mfhepp@gmail.com>, Dan Brickley <danbri@danbri.org>, ProjectParadigm-ICT-Program <metadataportals@yahoo.com>, Public-cogai <public-cogai@w3.org>, W3C AIKR CG <public-aikr@w3.org>, semantic-web <semantic-web@w3.org>
On 3/30/23 15:59, Adeel wrote: > You can't talk about regulation and compliance in this group, dan > doesn't like it as google doesn't care about those things. That is offensive. Please review W3C's Code of Ethics and Professional Conduct: https://www.w3.org/Consortium/cepc/ If you cannot contribute respectfully to the discussion, then please refrain from posting. Thanks, David Booth > > Thanks, > > Adeel > > On Thu, 30 Mar 2023 at 20:22, adasal <adam.saltiel@gmail.com > <mailto:adam.saltiel@gmail.com>> wrote: > > It's out of the bottle and will be played with. > > " .. being run on consumer laptops. And that’s not even thinking > about state level actors .. " > Large resources will be thrown at this. > > It was a long time ago that Henry Story (of course, many others too, > but more in this context) pointed out that, as to what pertains to > the truth, competing logical deductions cannot decide themselves. > > I just had this experience, and the details are not important. > > > The point is that, in this case, I asked the same question to GPT-4 > and perplexity.ai <http://perplexity.ai>, and they gave different > answers. > Since it was something I wanted to know the answer to, and it was > sufficiently complex, I was not in a position to judge which was > correct. > > Petitioning for funding for experts, i.e. researchers and university > professors. > Although it is absurd to think they would have time to mediate > between all the obscure information sorting correct from incorrect > and, of course, a person can be wrong too. > > Then there is the issue of attribution ... > At the moment, perplexity.ai <http://perplexity.ai> has a word salad > of dubious recent publications; GPT -4 has a "knowledge cutoff for > my training data is September 2021". It finds it difficult to reason > about time in any case, but these are details. > > Others in this email thread have cast doubt on Musk's motivation > (give it time to catch up) and Microsoft (didn't care for any > consequences by jumping in now). > > So there are issues of funding and control -- calling on the state > to intervene is appealing to the power next up the hierarchy, but > can such regulations be effective when administered by the state? > > That really just leaves us with grassroots education and everyday > intervention. > > Best on an important topic, > > > Adam > > Adam Saltiel > > > > On Wed, Mar 29, 2023 at 9:39 PM Martin Hepp <mfhepp@gmail.com > <mailto:mfhepp@gmail.com>> wrote: > > __ I could not agree more with Dan - a “non-proliferation” > agreement nor a moratorium of AI advancements is simply much > more unrealistic than it was with nukes. We hardly managed to > keep the number of crazy people with access to nukes under > control, but for building your next generation of AI, you will > not need anything but brain, programming skills, and commodity > resources. Machines will not take over humankind, but machines > can add giant levers to single individuals or groups. > > Best wishes > Martin > > --------------------------------------- > martin hepp > www: https://www.heppnetz.de/ <https://www.heppnetz.de/> > > >> On 29. Mar 2023, at 22:30, Dan Brickley <danbri@danbri.org >> <mailto:danbri@danbri.org>> wrote: >> >> >> >> On Wed, 29 Mar 2023 at 20:51, ProjectParadigm-ICT-Program >> <metadataportals@yahoo.com <mailto:metadataportals@yahoo.com>> >> wrote: >> >> This letter speaks for itself. >> >> https://www.reuters.com/technology/musk-experts-urge-pause-training-ai-systems-that-can-outperform-gpt-4-2023-03-29/ <https://www.reuters.com/technology/musk-experts-urge-pause-training-ai-systems-that-can-outperform-gpt-4-2023-03-29/> >> >> >> I may not want to put it as bluntly as Elon Musk, who >> cautioned against unregulated AI which he called "more >> dangerous than nukes", but when Nick Bostrom, the late >> Stephen Hawking, and dozens, no hundreds of international >> experts, scientists and industry leaders start ringing the >> bell, is is time to pause and reflect. >> >> Every aspect of daily life, every industry, education >> systems, academia and even our cognitive rights will be >> impacted. >> >> I would also like to point out that some science fiction >> authors have done a great job on very accurately >> predicting a dystopian future ruled by technology, perhaps >> the greatest of them all being Philip K. Dick. >> >> But there are dozens of other authors as well and they all >> give a fairly good impression what awaits us if we do not >> regulate and control the further development of AI now. >> >> >> I have a *lot* of worries, but the genie is out of the bottle. >> >> It’s 60 lines of code for the basics, >> https://jaykmody.com/blog/gpt-from-scratch/ >> <https://jaykmody.com/blog/gpt-from-scratch/> >> >> Facebook’s Llama model is out there, and being run on consumer >> laptops. And that’s not even thinking about state level >> actors, or how such regulation might be worded. >> >> For my part (and v personal opinion) I think focussing on >> education, sensible implementation guidelines, and trying to >> make sure the good outweighs the bad. >> >> Dan >> >> >> >> >> Milton Ponson >> GSM: +297 747 8280 >> PO Box 1154, Oranjestad >> Aruba, Dutch Caribbean >> Project Paradigm: Bringing the ICT tools for sustainable >> development to all stakeholders worldwide through >> collaborative research on applied mathematics, advanced >> modeling, software and standards development >>
Received on Thursday, 30 March 2023 20:14:53 UTC