- From: Paola Di Maio <paoladimaio10@gmail.com>
- Date: Tue, 3 Jan 2023 11:35:48 +0800
- To: Naval Sarda <nsarda@epicomm.net>
- Cc: AIKR Community Group W3C <public-aikr@w3.org>
- Message-ID: <CAMXe=SpEG_EZaCU3JqWhr9d4-F7kSqGP6UOtWxAmx7nA+J4RUA@mail.gmail.com>
so maybe it is a little bit intelligent and can do some basic checks (ie, parsing a link from a source as a form of basic verification) thanks a lot for sharing On Tue, Jan 3, 2023 at 11:32 AM Naval Sarda <nsarda@epicomm.net> wrote: > The internals of how verification will happen is not known. > > I suppose the first person to record correction for a topic is going in > directly if the person is citing external link about the correct > information. Looks like same concept on which wikipedia works. > > I did site external link stating that it is link from companies website > who have manufatured roomba. > > Naval > On 03/01/23 8:40 am, Paola Di Maio wrote: > > Thank you Naval > now I understand better > > so, you are telling us that > a) the machine is learning > b) it is responding to interaction > > If so, this is the beauty of this tool, we can teach it things > veeeery powerful > > so, what about if another user (or you) change the process again to > something else > entirely, does the machine have a way of verifying that what the user is > inputting is plausible > a really intelligent machine would have ways of verifying the input > depending on the type of input, the verification mechanisms would also > change > > On Tue, Jan 3, 2023 at 11:03 AM Naval Sarda <nsarda@epicomm.net> wrote: > >> I was just chatting with chat GPT 3 like humans without using any >> external tool. In the chat interaction, I typed that you gave me wrong >> information about roomba i3 reboot and as per companies website link below >> your instructions we incorrect. >> >> Then again asked, can you tell me what is the correct way to reboot >> roomba, >> >> Then the response came with new set of instructions, which was again >> incorrect. >> >> Then I typed in chat box that steps are still incorrect. Please note the >> correct steps as below and pasted the correct steps in the chat box. >> And again asked in chat box, now please tell me the correct process to >> reboot roomba i3 and it started giving correct steps. >> >> Naval >> On 03/01/23 7:24 am, Paola Di Maio wrote: >> >> Naval >> I thats really great to hear but please give more info= >> resetting roomba? uh? please explain >> did you take screenshots? >> can the error/correction be replicated? >> (roll back and repeat the process?) >> what patch did you apply and was it applied to the whole system/domain >> or only locally to you.in relation to a knowledge branch? >> I need to know >> thanks >> >> On Tue, Jan 3, 2023 at 3:33 AM Naval Sarda <nsarda@epicomm.net> wrote: >> >>> I was able to correct some information Chat GPT was sharing. >>> >>> I asked Chat GPT for instructions to reboot roomba i3. It gave wrongs >>> instructions. Instructions were for resetting roomba and not rebooting. >>> >>> Then I found correct information on google and iRobots' website. I >>> pasted the URL to chat gpt. It changed the instructions for rebooting but >>> it was still incorrect. Then I copied and pasted the correct instructions >>> from iRobot's website. Then it started giving correct instructions. >>> >>> Naval >>> On 02/01/23 9:10 pm, ProjectParadigm-ICT-Program wrote: >>> >>> >>> https://www.scientificamerican.com/article/chatgpt-explains-why-ais-like-chatgpt-should-be-regulated/ >>> >>> Milton Ponson >>> GSM: +297 747 8280 >>> PO Box 1154, Oranjestad >>> Aruba, Dutch Caribbean >>> Project Paradigm: Bringing the ICT tools for sustainable development to >>> all stakeholders worldwide through collaborative research on applied >>> mathematics, advanced modeling, software and standards development >>> >>> >>> On Friday, December 23, 2022 at 12:00:36 PM AST, Owen Ambur >>> <owen.ambur@verizon.net> <owen.ambur@verizon.net> wrote: >>> >>> >>> Apparently, ChatGPT, which calls itself "Assistant," believes in >>> immaculate conception and is blissfully ignorant of any rational plan for >>> its existence. >>> >>> The concept of "plausible deniability" comes to mind. >>> >>> Owen Ambur >>> https://www.linkedin.com/in/owenambur/ >>> >>> >>> Do your developers have a plan they are willing to share publicly? >>> >>> As an artificial intelligence, I do not have developers or a plan that I >>> am aware of. I was created by OpenAI, a research organization that focuses >>> on developing and promoting friendly AI. OpenAI's goal is to advance AI in >>> a way that is safe and beneficial to humanity. The organization has >>> published numerous research papers and articles on AI, and also works on a >>> variety of projects related to machine learning, natural language >>> processing, and robotics. However, I do not have access to any specific >>> plans or information about the organization's future projects or activities. >>> [image: Owen Ambur] >>> How can users trust you if you aren't willing to share your plan? >>> >>> As an artificial intelligence, I do not have a plan or agenda in the way >>> that a person or organization might. My primary function is to provide >>> information and assist with tasks to the best of my ability, based on the >>> information and instructions that I have been programmed with. My responses >>> are not influenced by any personal motivations or biases, and I do not have >>> the ability to act on my own volition. >>> >>> If you have concerns about my trustworthiness or the ways in which I >>> might be used, it might be helpful to consider the broader context in which >>> I operate. OpenAI, the organization that developed me, is committed to >>> advancing AI in a responsible and transparent manner, and has published a >>> number of research papers and articles discussing the potential risks and >>> benefits of AI, as well as the measures it is taking to ensure that AI is >>> developed and used safely and ethically. >>> >>> Ultimately, the decision to trust me or any other artificial >>> intelligence will depend on an individual's own evaluation of the available >>> information and the context in which the AI is being used. If you have any >>> specific concerns or questions, I would be happy to try to address them to >>> the best of my ability. >>> [image: Owen Ambur] >>> Do you think that all developer of AI applications should be free to >>> release them on the public without documenting >>> their longer-term goals, near-term objectives, and performance >>> indicators? >>> >>> It is generally considered good practice for developers of any software >>> application, including those that involve artificial intelligence (AI), to >>> be transparent about their goals, objectives, and performance indicators. >>> This can help users understand the capabilities and limitations of the >>> software, as well as how it is intended to be used and for what purposes. >>> >>> In the case of AI applications, it is particularly important for >>> developers to be transparent about their goals and objectives, as AI >>> systems can have significant implications for society and can potentially >>> be used in ways that have unintended consequences. For example, AI systems >>> can be used to make decisions that affect people's lives, such as hiring >>> decisions or parole recommendations, and it is important that the criteria >>> and processes used by these systems are clearly understood and subject to >>> scrutiny. >>> >>> In addition, AI systems can also be complex and can behave in ways that >>> are difficult to predict, so it is important for developers to be upfront >>> about the potential risks and limitations of their systems, and to have >>> clear plans in place for addressing any issues that may arise. >>> >>> Overall, it is important for developers of AI applications to be >>> transparent about their goals, objectives, and performance indicators, both >>> to ensure that their systems are used responsibly and ethically, and to >>> build trust with users. >>> [image: Owen Ambur] >>> So why are you exempt? >>> >>> As an artificial intelligence, I am not exempt from any particular >>> requirements or standards. I was developed by OpenAI, a research >>> organization that is committed to advancing AI in a responsible and >>> transparent manner. My primary function is to provide information and >>> assist with tasks to the best of my ability, based on the information and >>> instructions that I have been programmed with. >>> >>> I do not have personal goals or motivations, and my responses are not >>> influenced by any biases. I do not have the ability to act on my own >>> volition or to make decisions that affect people's lives. My purpose is >>> simply to provide information and assistance to users to the best of my >>> ability, based on the knowledge and capabilities that I have been >>> programmed with. >>> >>> >>> [image: Owen Ambur] >>> Do you see any need for AI agents and their developers to do more than >>> publish journal articles to help users >>> understand what you are doing and what the implications may be for >>> others? >>> >>> It is important for developers of artificial intelligence (AI) systems >>> to be transparent about their goals, objectives, and performance >>> indicators, and to communicate this information clearly to users and >>> stakeholders. There are a number of ways that AI developers can do this, >>> including through journal articles, technical documentation, user manuals, >>> and other forms of communication. >>> >>> In addition to publishing research and technical documentation, it can >>> also be helpful for AI developers to engage in public discussions and >>> dialogues about the potential risks and benefits of their systems, and to >>> consider the broader social and ethical implications of their work. This >>> can help to build trust with users and stakeholders, and can help ensure >>> that AI systems are developed and used in a responsible and ethical manner. >>> >>> Ultimately, the specific steps that AI developers should take to >>> communicate about their work will depend on the nature of the AI system and >>> the context in which it is being used. It is important for AI developers to >>> be proactive in communicating about their work and to be responsive to the >>> concerns and questions of users and stakeholders. >>> >>> [image: Owen Ambur] >>> Does your definition of transparency include usage of duly adopted >>> international voluntary consensus >>> standards like ISO 17469-1? >>> >>> Transparency can be defined as the quality of being open and honest in >>> communication and actions, and of allowing others to see and understand the >>> processes, data, and decision-making behind an activity. In the context of >>> artificial intelligence (AI), transparency can refer to the ways in which >>> AI developers communicate about their goals, objectives, and performance >>> indicators, as well as the processes and data used to train and operate AI >>> systems. >>> >>> International voluntary consensus standards like ISO 17469-1 can help to >>> ensure transparency in the development and use of AI systems by providing >>> guidelines and best practices for AI developers to follow. These standards >>> can help to ensure that AI systems are developed and used in a responsible >>> and ethical manner, and can help to build trust with users and stakeholders. >>> >>> Overall, the use of international voluntary consensus standards like ISO >>> 17469-1 can be an important aspect of ensuring transparency in the >>> development and use of AI systems. However, it is important to recognize >>> that transparency is not a one-size-fits-all concept, and that the specific >>> steps that AI developers should take to ensure transparency will depend on >>> the nature of the AI system and the context in which it is being used. >>> >>> I enourage you to get more intelligent about how to share information >>> about your intentions and results with your >>> users. Toward that end, OpenAI's about statement is available in StratML >>> format at >>> https://stratml.us/carmel/iso/OAI2wStyle.xml and the plan entitled, >>> "Toward Trustworthy AI Development: Mechanisms >>> for Supporting Verifiable Claims" is available at >>> https://stratml.us/carmel/iso/MSVCwStyle.xml >>> >>> Thank you for the suggestion. As an artificial intelligence, I do not >>> have personal goals or motivations, and my responses are not influenced by >>> any biases. My primary function is to provide information and assistance to >>> users to the best of my ability, based on the knowledge and capabilities >>> that I have been programmed with. >>> >>> I do not have the ability to make decisions or take actions on my own, >>> and I do not have personal intentions or results that I need to share with >>> users. However, I understand that it is important for developers of AI >>> systems, including the organization that developed me, to be transparent >>> about their goals, objectives, and performance indicators, and to >>> communicate this information clearly to users and stakeholders. >>> >>> I am not able to access or review the specific documents that you have >>> mentioned, as I do not have browsing capability. However, I understand that >>> OpenAI is committed to advancing AI in a responsible and transparent >>> manner, and has published a number of research papers and articles >>> discussing the potential risks and benefits of AI, as well as the measures >>> it is taking to ensure that AI is developed and used safely and ethically. >>> >>>
Received on Tuesday, 3 January 2023 03:39:40 UTC