- From: Naval Sarda <nsarda@epicomm.net>
- Date: Tue, 3 Jan 2023 09:01:39 +0530
- To: Paola Di Maio <paoladimaio10@gmail.com>
- Cc: AIKR Community Group W3C <public-aikr@w3.org>
- Message-ID: <0204c384-846e-17e9-ac0b-b2c37a1a5931@epicomm.net>
The internals of how verification will happen is not known. I suppose the first person to record correction for a topic is going in directly if the person is citing external link about the correct information. Looks like same concept on which wikipedia works. I did site external link stating that it is link from companies website who have manufatured roomba. Naval On 03/01/23 8:40 am, Paola Di Maio wrote: > Thank you Naval > now I understand better > > so, you are telling us that > a) the machine is learning > b) it is responding to interaction > > If so, this is the beauty of this tool, we can teach it things > veeeery powerful > > so, what about if another user (or you) change the process again to > something else > entirely, does the machine have a way of verifying that what the user > is inputting is plausible > a really intelligent machine would have ways of verifying the input > depending on the type of input, the verification mechanisms would also > change > > On Tue, Jan 3, 2023 at 11:03 AM Naval Sarda <nsarda@epicomm.net> wrote: > > I was just chatting with chat GPT 3 like humans without using any > external tool. In the chat interaction, I typed that you gave me > wrong information about roomba i3 reboot and as per companies > website link below your instructions we incorrect. > > Then again asked, can you tell me what is the correct way to > reboot roomba, > > Then the response came with new set of instructions, which was > again incorrect. > > Then I typed in chat box that steps are still incorrect. Please > note the correct steps as below and pasted the correct steps in > the chat box. > And again asked in chat box, now please tell me the correct > process to reboot roomba i3 and it started giving correct steps. > > Naval > > On 03/01/23 7:24 am, Paola Di Maio wrote: >> Naval >> I thats really great to hear but please give more info= >> resetting roomba? uh? please explain >> did you take screenshots? >> can the error/correction be replicated? >> (roll back and repeat the process?) >> what patch did you apply and was it applied to the whole >> system/domain >> or only locally to you.in <http://you.in> relation to a knowledge >> branch? >> I need to know >> thanks >> >> On Tue, Jan 3, 2023 at 3:33 AM Naval Sarda <nsarda@epicomm.net> >> wrote: >> >> I was able to correct some information Chat GPT was sharing. >> >> I asked Chat GPT for instructions to reboot roomba i3. It >> gave wrongs instructions. Instructions were for resetting >> roomba and not rebooting. >> >> Then I found correct information on google and iRobots' >> website. I pasted the URL to chat gpt. It changed the >> instructions for rebooting but it was still incorrect. Then I >> copied and pasted the correct instructions from iRobot's >> website. Then it started giving correct instructions. >> >> Naval >> >> On 02/01/23 9:10 pm, ProjectParadigm-ICT-Program wrote: >>> https://www.scientificamerican.com/article/chatgpt-explains-why-ais-like-chatgpt-should-be-regulated/ >>> >>> Milton Ponson >>> GSM: +297 747 8280 >>> PO Box 1154, Oranjestad >>> Aruba, Dutch Caribbean >>> Project Paradigm: Bringing the ICT tools for sustainable >>> development to all stakeholders worldwide through >>> collaborative research on applied mathematics, advanced >>> modeling, software and standards development >>> >>> >>> On Friday, December 23, 2022 at 12:00:36 PM AST, Owen Ambur >>> <owen.ambur@verizon.net> <mailto:owen.ambur@verizon.net> wrote: >>> >>> >>> Apparently, ChatGPT, which calls itself "Assistant," >>> believes in immaculate conception and is blissfully ignorant >>> of any rational plan for its existence. >>> >>> The concept of "plausible deniability" comes to mind. >>> >>> Owen Ambur >>> https://www.linkedin.com/in/owenambur/ >>> >>> >>> Do your developers have a plan they are willing to share >>> publicly? >>> >>> As an artificial intelligence, I do not have developers or a >>> plan that I am aware of. I was created by OpenAI, a research >>> organization that focuses on developing and promoting >>> friendly AI. OpenAI's goal is to advance AI in a way that is >>> safe and beneficial to humanity. The organization has >>> published numerous research papers and articles on AI, and >>> also works on a variety of projects related to machine >>> learning, natural language processing, and robotics. >>> However, I do not have access to any specific plans or >>> information about the organization's future projects or >>> activities. >>> >>> Owen Ambur >>> How can users trust you if you aren't willing to share your >>> plan? >>> >>> As an artificial intelligence, I do not have a plan or >>> agenda in the way that a person or organization might. My >>> primary function is to provide information and assist with >>> tasks to the best of my ability, based on the information >>> and instructions that I have been programmed with. My >>> responses are not influenced by any personal motivations or >>> biases, and I do not have the ability to act on my own volition. >>> >>> If you have concerns about my trustworthiness or the ways in >>> which I might be used, it might be helpful to consider the >>> broader context in which I operate. OpenAI, the organization >>> that developed me, is committed to advancing AI in a >>> responsible and transparent manner, and has published a >>> number of research papers and articles discussing the >>> potential risks and benefits of AI, as well as the measures >>> it is taking to ensure that AI is developed and used safely >>> and ethically. >>> >>> Ultimately, the decision to trust me or any other artificial >>> intelligence will depend on an individual's own evaluation >>> of the available information and the context in which the AI >>> is being used. If you have any specific concerns or >>> questions, I would be happy to try to address them to the >>> best of my ability. >>> >>> Owen Ambur >>> Do you think that all developer of AI applications should be >>> free to release them on the public without documenting >>> their longer-term goals, near-term objectives, and >>> performance indicators? >>> >>> It is generally considered good practice for developers of >>> any software application, including those that involve >>> artificial intelligence (AI), to be transparent about their >>> goals, objectives, and performance indicators. This can help >>> users understand the capabilities and limitations of the >>> software, as well as how it is intended to be used and for >>> what purposes. >>> >>> In the case of AI applications, it is particularly important >>> for developers to be transparent about their goals and >>> objectives, as AI systems can have significant implications >>> for society and can potentially be used in ways that have >>> unintended consequences. For example, AI systems can be used >>> to make decisions that affect people's lives, such as hiring >>> decisions or parole recommendations, and it is important >>> that the criteria and processes used by these systems are >>> clearly understood and subject to scrutiny. >>> >>> In addition, AI systems can also be complex and can behave >>> in ways that are difficult to predict, so it is important >>> for developers to be upfront about the potential risks and >>> limitations of their systems, and to have clear plans in >>> place for addressing any issues that may arise. >>> >>> Overall, it is important for developers of AI applications >>> to be transparent about their goals, objectives, and >>> performance indicators, both to ensure that their systems >>> are used responsibly and ethically, and to build trust with >>> users. >>> >>> Owen Ambur >>> So why are you exempt? >>> >>> As an artificial intelligence, I am not exempt from any >>> particular requirements or standards. I was developed by >>> OpenAI, a research organization that is committed to >>> advancing AI in a responsible and transparent manner. My >>> primary function is to provide information and assist with >>> tasks to the best of my ability, based on the information >>> and instructions that I have been programmed with. >>> >>> I do not have personal goals or motivations, and my >>> responses are not influenced by any biases. I do not have >>> the ability to act on my own volition or to make decisions >>> that affect people's lives. My purpose is simply to provide >>> information and assistance to users to the best of my >>> ability, based on the knowledge and capabilities that I have >>> been programmed with. >>> >>> >>> >>> Owen Ambur >>> Do you see any need for AI agents and their developers to do >>> more than publish journal articles to help users >>> understand what you are doing and what the implications may >>> be for others? >>> >>> It is important for developers of artificial intelligence >>> (AI) systems to be transparent about their goals, >>> objectives, and performance indicators, and to communicate >>> this information clearly to users and stakeholders. There >>> are a number of ways that AI developers can do this, >>> including through journal articles, technical documentation, >>> user manuals, and other forms of communication. >>> >>> In addition to publishing research and technical >>> documentation, it can also be helpful for AI developers to >>> engage in public discussions and dialogues about the >>> potential risks and benefits of their systems, and to >>> consider the broader social and ethical implications of >>> their work. This can help to build trust with users and >>> stakeholders, and can help ensure that AI systems are >>> developed and used in a responsible and ethical manner. >>> >>> Ultimately, the specific steps that AI developers should >>> take to communicate about their work will depend on the >>> nature of the AI system and the context in which it is being >>> used. It is important for AI developers to be proactive in >>> communicating about their work and to be responsive to the >>> concerns and questions of users and stakeholders. >>> >>> >>> Owen Ambur >>> Does your definition of transparency include usage of duly >>> adopted international voluntary consensus >>> standards like ISO 17469-1? >>> >>> Transparency can be defined as the quality of being open and >>> honest in communication and actions, and of allowing others >>> to see and understand the processes, data, and >>> decision-making behind an activity. In the context of >>> artificial intelligence (AI), transparency can refer to the >>> ways in which AI developers communicate about their goals, >>> objectives, and performance indicators, as well as the >>> processes and data used to train and operate AI systems. >>> >>> International voluntary consensus standards like ISO 17469-1 >>> can help to ensure transparency in the development and use >>> of AI systems by providing guidelines and best practices for >>> AI developers to follow. These standards can help to ensure >>> that AI systems are developed and used in a responsible and >>> ethical manner, and can help to build trust with users and >>> stakeholders. >>> >>> Overall, the use of international voluntary consensus >>> standards like ISO 17469-1 can be an important aspect of >>> ensuring transparency in the development and use of AI >>> systems. However, it is important to recognize that >>> transparency is not a one-size-fits-all concept, and that >>> the specific steps that AI developers should take to ensure >>> transparency will depend on the nature of the AI system and >>> the context in which it is being used. >>> >>> >>> I enourage you to get more intelligent about how to share >>> information about your intentions and results with your >>> users. Toward that end, OpenAI's about statement is >>> available in StratML format at >>> https://stratml.us/carmel/iso/OAI2wStyle.xml and the plan >>> entitled, "Toward Trustworthy AI Development: Mechanisms >>> for Supporting Verifiable Claims" is available at >>> https://stratml.us/carmel/iso/MSVCwStyle.xml >>> >>> Thank you for the suggestion. As an artificial intelligence, >>> I do not have personal goals or motivations, and my >>> responses are not influenced by any biases. My primary >>> function is to provide information and assistance to users >>> to the best of my ability, based on the knowledge and >>> capabilities that I have been programmed with. >>> >>> I do not have the ability to make decisions or take actions >>> on my own, and I do not have personal intentions or results >>> that I need to share with users. However, I understand that >>> it is important for developers of AI systems, including the >>> organization that developed me, to be transparent about >>> their goals, objectives, and performance indicators, and to >>> communicate this information clearly to users and stakeholders. >>> >>> I am not able to access or review the specific documents >>> that you have mentioned, as I do not have browsing >>> capability. However, I understand that OpenAI is committed >>> to advancing AI in a responsible and transparent manner, and >>> has published a number of research papers and articles >>> discussing the potential risks and benefits of AI, as well >>> as the measures it is taking to ensure that AI is developed >>> and used safely and ethically. >>> >>>
Received on Tuesday, 3 January 2023 03:32:34 UTC