- From: Adeel <aahmad1811@gmail.com>
- Date: Mon, 3 Apr 2023 00:28:06 +0100
- To: Melvin Carvalho <melvincarvalho@gmail.com>
- Cc: "Patrick J. Hayes" <phayes@ihmc.org>, Dan Brickley <danbri@danbri.org>, Martin Hepp <mfhepp@gmail.com>, ProjectParadigm-ICT-Program <metadataportals@yahoo.com>, Public-cogai <public-cogai@w3.org>, W3C AIKR CG <public-aikr@w3.org>, adasal <adam.saltiel@gmail.com>, semantic-web <semantic-web@w3.org>
- Message-ID: <CALpEXW09+i7OK6O2V1QekfDkgA3Tc7=9byqUV62VUe6yUOxCvQ@mail.gmail.com>
so, you fed PII data to chatgpt, proving your questionable ethics… On Sun, 2 Apr 2023 at 03:15, Melvin Carvalho <melvincarvalho@gmail.com> wrote: > > > so 1. 4. 2023 v 23:29 odesílatel Adeel <aahmad1811@gmail.com> napsal: > >> It is a constructive discussion on the general ethics - it is a core >> competency in every professionalism. Not left to legal speculation whether >> it be the use of transformer models, code of conduct, or selectively >> defined for others, practice what you preach which is the real issue with >> ethics these days. >> > > Perhaps calm down a bit and stay on-topic > > As an interesting exercise I ran this tread through chatGPT asking: "rate > this statement as whether or not it is passive agressive on a scale of 1 to > 10 and why" -- give it a try if you want to see some interesting results > > What is came up with passive aggression: > > Hugh - 3 > Dan - 5 > Pat - 4 > Adeel - 5,7 > Nathan (control) - 2 > > Perhaps if run again it would give different numbers. I expect AI is > going to be integrated into all sorts of tools, so that we'll get flags and > pointers in email before sending. Which will change the nature of mail > lists forever! > > As a side note I chat to new developers about semantic web all the time. > One recently said to me: "I hope stuff like this can popularize linked data > without the smugness". > > This is actually one of the more pleasant, jolly threads. But we probably > can and will do better as a list, when it starts to be common place for AI > tools to help us. > > It is a technology that huge implications for the semantic web. Because > the semantic web is about meaning. There are times when technology moves > too fast, and has societal implications, so imho, the question is legit. > Indeed in the AI world this topic has been the big one in the last week, > with dozens of different views. It's interesting. Emad has in any case > said he thinks gpt v.next will take 6-9 months to train anyway. > > The main limitation I have found is that it can do small tasks reasonably > well, but when you hit the 32k limit, it struggles just as much as we all > do with larger topics or code bases. Perhaps that will chain soon with > plugins tho, we'll see! > > >> >> On Sat, 1 Apr 2023 at 22:12, Patrick J. Hayes <phayes@ihmc.org> wrote: >> >>> Adeel, greetings. >>> >>> The code of conduct applies to us all, including of course Dan B. But >>> the point here is that Dan has not violated that code, whereas you have, >>> first by directly insulting Dan by implying that his posts are dishonest or >>> motivated by dark or sinister commercial forces, and then by continuing >>> this thread of thinly veiled hostility when there is no reason for it. I am >>> not an administrator of this group, but if I were you would now be banned >>> from it. >>> >>> Please find some more constructive way to contribute to the discussion. >>> >>> Pat Hayes >>> >>> On Apr 1, 2023, at 1:22 PM, Adeel <aahmad1811@gmail.com> wrote: >>> >>> >>> Hello, >>> >>> That sounds like a legal speculation, or do you only selectively >>> discriminate on group members and their backgrounds when you point that out? >>> Does the W3C code of conduct not apply to you after 25+ years of being >>> here? >>> >>> Thanks, >>> >>> Adeel >>> >>> >>> On Thu, 30 Mar 2023 at 23:33, Dan Brickley <danbri@danbri.org> wrote: >>> >>>> On Thu, 30 Mar 2023 at 20:59, Adeel <aahmad1811@gmail.com> wrote: >>>> >>>>> Hello, >>>>> >>>>> You can't talk about regulation and compliance in this group, dan >>>>> doesn't like it as google doesn't care about those things. >>>>> >>>> >>>> Rather, Google doesn’t delegate to me any permission to speak on its >>>> behalf on these matters, unsurprisingly enough. Google is also and >>>> organization of many thousand employees, with quite some range of views. I >>>> choose not to share several of mine here right now, but I am broadly in >>>> favour of sensible laws where they have some chance of working. >>>> >>>> As in many situations it sometimes makes sense to treat a company as a >>>> kind of pretend person, and sometimes to look at it as a more complex >>>> system of forces and interests. Skim >>>> >>>> https://www.wired.com/story/inside-google-three-years-misery-happiest-company-tech/ >>>> <https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.wired.com%2Fstory%2Finside-google-three-years-misery-happiest-company-tech%2F&data=05%7C01%7Cphayes%40ihmc.us%7Cd884dfb499ef423f124f08db32ef753e%7C2b38115bebad4aba9ea3b3779d8f4f43%7C1%7C0%7C638159776297688875%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=akFFkIcIXEgvQWjOdtQO4QjZOzAjTOJ82zH1mJwAb%2Bo%3D&reserved=0> >>>> to see some of what bubbles beneath the surface. >>>> >>>> This email list is just an email list. W3C no longer accords it >>>> “interest group“ status, as it did from 1999 when I migrated the older >>>> RDF-DEV list to W3C to form the RDF Interest Group on www-rdf-interest@ >>>> list. W3C doesn’t officially listen here even on RDF topics, let alone >>>> areas like modern AI and their social impact which aren’t obviously our >>>> core competency. >>>> >>>> Having been here 25+ years I have some instincts about which topics >>>> will just fill up email inboxes with no ultimate impact on the world and >>>> benefit. Right now “something must be banned” threads on AI look to me to >>>> fall in that category. >>>> >>>> Cheers, >>>> >>>> Dan >>>> >>> >>> >>>> >>>> Thanks, >>>>> >>>>> Adeel >>>>> >>>>> On Thu, 30 Mar 2023 at 20:22, adasal <adam.saltiel@gmail.com> wrote: >>>>> >>>>>> It's out of the bottle and will be played with. >>>>>> >>>>>> " .. being run on consumer laptops. And that’s not even thinking >>>>>> about state level actors .. " >>>>>> Large resources will be thrown at this. >>>>>> >>>>>> It was a long time ago that Henry Story (of course, many others too, >>>>>> but more in this context) pointed out that, as to what pertains to the >>>>>> truth, competing logical deductions cannot decide themselves. >>>>>> >>>>>> I just had this experience, and the details are not important. >>>>>> >>>>>> >>>>>> The point is that, in this case, I asked the same question to GPT-4 >>>>>> and perplexity.ai >>>>>> <https://nam02.safelinks.protection.outlook.com/?url=http%3A%2F%2Fperplexity.ai%2F&data=05%7C01%7Cphayes%40ihmc.us%7Cd884dfb499ef423f124f08db32ef753e%7C2b38115bebad4aba9ea3b3779d8f4f43%7C1%7C0%7C638159776297688875%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=twBLGhgQjubMJb6eq223Hsx8kKJMs8Tiq7jMBfpFyPc%3D&reserved=0>, >>>>>> and they gave different answers. >>>>>> Since it was something I wanted to know the answer to, and it was >>>>>> sufficiently complex, I was not in a position to judge which was correct. >>>>>> >>>>>> Petitioning for funding for experts, i.e. researchers and university >>>>>> professors. >>>>>> Although it is absurd to think they would have time to mediate >>>>>> between all the obscure information sorting correct from incorrect and, of >>>>>> course, a person can be wrong too. >>>>>> >>>>>> Then there is the issue of attribution ... >>>>>> At the moment, perplexity.ai >>>>>> <https://nam02.safelinks.protection.outlook.com/?url=http%3A%2F%2Fperplexity.ai%2F&data=05%7C01%7Cphayes%40ihmc.us%7Cd884dfb499ef423f124f08db32ef753e%7C2b38115bebad4aba9ea3b3779d8f4f43%7C1%7C0%7C638159776297688875%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=twBLGhgQjubMJb6eq223Hsx8kKJMs8Tiq7jMBfpFyPc%3D&reserved=0> >>>>>> has a word salad of dubious recent publications; GPT -4 has a "knowledge >>>>>> cutoff for my training data is September 2021". It finds it difficult to >>>>>> reason about time in any case, but these are details. >>>>>> >>>>>> Others in this email thread have cast doubt on Musk's motivation >>>>>> (give it time to catch up) and Microsoft (didn't care for any consequences >>>>>> by jumping in now). >>>>>> >>>>>> So there are issues of funding and control -- calling on the state to >>>>>> intervene is appealing to the power next up the hierarchy, but can such >>>>>> regulations be effective when administered by the state? >>>>>> >>>>>> That really just leaves us with grassroots education and everyday >>>>>> intervention. >>>>>> >>>>>> Best on an important topic, >>>>>> >>>>>> >>>>>> Adam >>>>>> >>>>>> Adam Saltiel >>>>>> >>>>>> >>>>>> >>>>>> On Wed, Mar 29, 2023 at 9:39 PM Martin Hepp <mfhepp@gmail.com> wrote: >>>>>> >>>>>>> I could not agree more with Dan - a “non-proliferation” agreement >>>>>>> nor a moratorium of AI advancements is simply much more unrealistic than it >>>>>>> was with nukes. We hardly managed to keep the number of crazy people with >>>>>>> access to nukes under control, but for building your next generation of AI, >>>>>>> you will not need anything but brain, programming skills, and commodity >>>>>>> resources. Machines will not take over humankind, but machines can add >>>>>>> giant levers to single individuals or groups. >>>>>>> >>>>>>> Best wishes >>>>>>> Martin >>>>>>> >>>>>>> --------------------------------------- >>>>>>> martin hepp >>>>>>> www: https://www.heppnetz.de/ >>>>>>> <https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.heppnetz.de%2F&data=05%7C01%7Cphayes%40ihmc.us%7Cd884dfb499ef423f124f08db32ef753e%7C2b38115bebad4aba9ea3b3779d8f4f43%7C1%7C0%7C638159776297845108%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=FCqp2o420lD%2FDJA%2BsW2xmUV1SL4eb4FqubE%2FFbQiwv0%3D&reserved=0> >>>>>>> >>>>>>> >>>>>>> On 29. Mar 2023, at 22:30, Dan Brickley <danbri@danbri.org> wrote: >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Wed, 29 Mar 2023 at 20:51, ProjectParadigm-ICT-Program < >>>>>>> metadataportals@yahoo.com> wrote: >>>>>>> >>>>>>> This letter speaks for itself. >>>>>>> >>>>>>> >>>>>>> https://www.reuters.com/technology/musk-experts-urge-pause-training-ai-systems-that-can-outperform-gpt-4-2023-03-29/ >>>>>>> <https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.reuters.com%2Ftechnology%2Fmusk-experts-urge-pause-training-ai-systems-that-can-outperform-gpt-4-2023-03-29%2F&data=05%7C01%7Cphayes%40ihmc.us%7Cd884dfb499ef423f124f08db32ef753e%7C2b38115bebad4aba9ea3b3779d8f4f43%7C1%7C0%7C638159776297845108%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=n%2BG6aomh5Ad6qwdk9euMYmYvQFLs0dSYFMpwJ%2FquyPo%3D&reserved=0> >>>>>>> >>>>>>> >>>>>>> I may not want to put it as bluntly as Elon Musk, who cautioned >>>>>>> against unregulated AI which he called "more dangerous than nukes", but >>>>>>> when Nick Bostrom, the late Stephen Hawking, and dozens, no hundreds of >>>>>>> international experts, scientists and industry leaders start ringing the >>>>>>> bell, is is time to pause and reflect. >>>>>>> >>>>>>> Every aspect of daily life, every industry, education systems, >>>>>>> academia and even our cognitive rights will be impacted. >>>>>>> >>>>>>> I would also like to point out that some science fiction authors >>>>>>> have done a great job on very accurately predicting a dystopian future >>>>>>> ruled by technology, perhaps the greatest of them all being Philip K. Dick. >>>>>>> >>>>>>> But there are dozens of other authors as well and they all give a >>>>>>> fairly good impression what awaits us if we do not regulate and control the >>>>>>> further development of AI now. >>>>>>> >>>>>>> >>>>>>> I have a *lot* of worries, but the genie is out of the bottle. >>>>>>> >>>>>>> It’s 60 lines of code for the basics, >>>>>>> https://jaykmody.com/blog/gpt-from-scratch/ >>>>>>> <https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fjaykmody.com%2Fblog%2Fgpt-from-scratch%2F&data=05%7C01%7Cphayes%40ihmc.us%7Cd884dfb499ef423f124f08db32ef753e%7C2b38115bebad4aba9ea3b3779d8f4f43%7C1%7C0%7C638159776297845108%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=2QEWpct5NS36zpwh2zyEISddXRBZMJeJtvMcQSJ8HME%3D&reserved=0> >>>>>>> >>>>>>> Facebook’s Llama model is out there, and being run on consumer >>>>>>> laptops. And that’s not even thinking about state level actors, or how such >>>>>>> regulation might be worded. >>>>>>> >>>>>>> For my part (and v personal opinion) I think focussing on education, >>>>>>> sensible implementation guidelines, and trying to make sure the good >>>>>>> outweighs the bad. >>>>>>> >>>>>>> Dan >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> Milton Ponson >>>>>>> GSM: +297 747 8280 >>>>>>> PO Box 1154, Oranjestad >>>>>>> Aruba, Dutch Caribbean >>>>>>> Project Paradigm: Bringing the ICT tools for sustainable >>>>>>> development to all stakeholders worldwide through collaborative research on >>>>>>> applied mathematics, advanced modeling, software and standards development >>>>>>> >>>>>>> >>>
Received on Sunday, 2 April 2023 23:28:32 UTC