- From: Mike Bergman <mike@mkbergman.com>
- Date: Mon, 2 Jan 2023 22:25:59 -0600
- To: Paola Di Maio <paoladimaio10@gmail.com>, Naval Sarda <nsarda@epicomm.net>
- Cc: AIKR Community Group W3C <public-aikr@w3.org>
- Message-ID: <304dea3a-e24a-7be6-b3c2-f582d3922b4b@mkbergman.com>
Hi Paola, I think this is state-of-the-art for textual Q & A systems. To learn further, it is likely essential to sign up for a free account <https://chat.openai.com/auth/login> (for now) and begin to play. To narrow or expand queries, think of your questions to the system as Venn phrases in a natural language construction with the standard AND or OR Boolean operators. To exclude things, you can say natural language things like, "Besides these [A, B, or C conditions]' or you can specify the relations such as 'similar' or 'different' or add the action Ws (who, what, where, when, why), etc. I personally always ask the system in a polite and respectful way. But I suspect the system does not require it; one could be demanding or snarky if that is your preferred style. The system apparently learns (short term? earlier threads), but I find comparing the gaps in my expected answers to my questions to be most helpful in refining my next query. I find the system to be an assistant, one who finds more connections or relations or helps one remember prior known connections. It is not infallible and misses many connections. But notice I said 'one' there. That is because, if not intelligent, the system is a knowledgeable and interactive agent. I don't know as we need to get more cosmic than that. Knowledge nuggets, yes; new insights, no. Still, pretty helpful in the right contexts. Best, Mike On 1/2/2023 9:35 PM, Paola Di Maio wrote: > so maybe it is a little bit intelligent and can do some basic checks > (ie, parsing a link from a source as a form of basic verification) > > thanks a lot for sharing > > On Tue, Jan 3, 2023 at 11:32 AM Naval Sarda <nsarda@epicomm.net> wrote: > > The internals of how verification will happen is not known. > > I suppose the first person to record correction for a topic is > going in directly if the person is citing external link about the > correct information. Looks like same concept on which wikipedia works. > > I did site external link stating that it is link from companies > website who have manufatured roomba. > > Naval > > On 03/01/23 8:40 am, Paola Di Maio wrote: >> Thank you Naval >> now I understand better >> >> so, you are telling us that >> a) the machine is learning >> b) it is responding to interaction >> >> If so, this is the beauty of this tool, we can teach it things >> veeeery powerful >> >> so, what about if another user (or you) change the process again >> to something else >> entirely, does the machine have a way of verifying that what the >> user is inputting is plausible >> a really intelligent machine would have ways of verifying the input >> depending on the type of input, the verification mechanisms would >> also change >> >> On Tue, Jan 3, 2023 at 11:03 AM Naval Sarda <nsarda@epicomm.net> >> wrote: >> >> I was just chatting with chat GPT 3 like humans without using >> any external tool. In the chat interaction, I typed that you >> gave me wrong information about roomba i3 reboot and as per >> companies website link below your instructions we incorrect. >> >> Then again asked, can you tell me what is the correct way to >> reboot roomba, >> >> Then the response came with new set of instructions, which >> was again incorrect. >> >> Then I typed in chat box that steps are still incorrect. >> Please note the correct steps as below and pasted the correct >> steps in the chat box. >> And again asked in chat box, now please tell me the correct >> process to reboot roomba i3 and it started giving correct steps. >> >> Naval >> >> On 03/01/23 7:24 am, Paola Di Maio wrote: >>> Naval >>> I thats really great to hear but please give more info= >>> resetting roomba? uh? please explain >>> did you take screenshots? >>> can the error/correction be replicated? >>> (roll back and repeat the process?) >>> what patch did you apply and was it applied to the whole >>> system/domain >>> or only locally to you.in <http://you.in> relation to a >>> knowledge branch? >>> I need to know >>> thanks >>> >>> On Tue, Jan 3, 2023 at 3:33 AM Naval Sarda >>> <nsarda@epicomm.net> wrote: >>> >>> I was able to correct some information Chat GPT was sharing. >>> >>> I asked Chat GPT for instructions to reboot roomba i3. >>> It gave wrongs instructions. Instructions were for >>> resetting roomba and not rebooting. >>> >>> Then I found correct information on google and iRobots' >>> website. I pasted the URL to chat gpt. It changed the >>> instructions for rebooting but it was still incorrect. >>> Then I copied and pasted the correct instructions from >>> iRobot's website. Then it started giving correct >>> instructions. >>> >>> Naval >>> >>> On 02/01/23 9:10 pm, ProjectParadigm-ICT-Program wrote: >>>> https://www.scientificamerican.com/article/chatgpt-explains-why-ais-like-chatgpt-should-be-regulated/ >>>> >>>> Milton Ponson >>>> GSM: +297 747 8280 >>>> PO Box 1154, Oranjestad >>>> Aruba, Dutch Caribbean >>>> Project Paradigm: Bringing the ICT tools for >>>> sustainable development to all stakeholders worldwide >>>> through collaborative research on applied mathematics, >>>> advanced modeling, software and standards development >>>> >>>> >>>> On Friday, December 23, 2022 at 12:00:36 PM AST, Owen >>>> Ambur <owen.ambur@verizon.net> >>>> <mailto:owen.ambur@verizon.net> wrote: >>>> >>>> >>>> Apparently, ChatGPT, which calls itself "Assistant," >>>> believes in immaculate conception and is blissfully >>>> ignorant of any rational plan for its existence. >>>> >>>> The concept of "plausible deniability" comes to mind. >>>> >>>> Owen Ambur >>>> https://www.linkedin.com/in/owenambur/ >>>> >>>> >>>> Do your developers have a plan they are willing to >>>> share publicly? >>>> >>>> As an artificial intelligence, I do not have developers >>>> or a plan that I am aware of. I was created by OpenAI, >>>> a research organization that focuses on developing and >>>> promoting friendly AI. OpenAI's goal is to advance AI >>>> in a way that is safe and beneficial to humanity. The >>>> organization has published numerous research papers and >>>> articles on AI, and also works on a variety of projects >>>> related to machine learning, natural language >>>> processing, and robotics. However, I do not have access >>>> to any specific plans or information about the >>>> organization's future projects or activities. >>>> >>>> Owen Ambur >>>> How can users trust you if you aren't willing to share >>>> your plan? >>>> >>>> As an artificial intelligence, I do not have a plan or >>>> agenda in the way that a person or organization might. >>>> My primary function is to provide information and >>>> assist with tasks to the best of my ability, based on >>>> the information and instructions that I have been >>>> programmed with. My responses are not influenced by any >>>> personal motivations or biases, and I do not have the >>>> ability to act on my own volition. >>>> >>>> If you have concerns about my trustworthiness or the >>>> ways in which I might be used, it might be helpful to >>>> consider the broader context in which I operate. >>>> OpenAI, the organization that developed me, is >>>> committed to advancing AI in a responsible and >>>> transparent manner, and has published a number of >>>> research papers and articles discussing the potential >>>> risks and benefits of AI, as well as the measures it is >>>> taking to ensure that AI is developed and used safely >>>> and ethically. >>>> >>>> Ultimately, the decision to trust me or any other >>>> artificial intelligence will depend on an individual's >>>> own evaluation of the available information and the >>>> context in which the AI is being used. If you have any >>>> specific concerns or questions, I would be happy to try >>>> to address them to the best of my ability. >>>> >>>> Owen Ambur >>>> Do you think that all developer of AI applications >>>> should be free to release them on the public without >>>> documenting >>>> their longer-term goals, near-term objectives, and >>>> performance indicators? >>>> >>>> It is generally considered good practice for developers >>>> of any software application, including those that >>>> involve artificial intelligence (AI), to be transparent >>>> about their goals, objectives, and performance >>>> indicators. This can help users understand the >>>> capabilities and limitations of the software, as well >>>> as how it is intended to be used and for what purposes. >>>> >>>> In the case of AI applications, it is particularly >>>> important for developers to be transparent about their >>>> goals and objectives, as AI systems can have >>>> significant implications for society and can >>>> potentially be used in ways that have unintended >>>> consequences. For example, AI systems can be used to >>>> make decisions that affect people's lives, such as >>>> hiring decisions or parole recommendations, and it is >>>> important that the criteria and processes used by these >>>> systems are clearly understood and subject to scrutiny. >>>> >>>> In addition, AI systems can also be complex and can >>>> behave in ways that are difficult to predict, so it is >>>> important for developers to be upfront about the >>>> potential risks and limitations of their systems, and >>>> to have clear plans in place for addressing any issues >>>> that may arise. >>>> >>>> Overall, it is important for developers of AI >>>> applications to be transparent about their goals, >>>> objectives, and performance indicators, both to ensure >>>> that their systems are used responsibly and ethically, >>>> and to build trust with users. >>>> >>>> Owen Ambur >>>> So why are you exempt? >>>> >>>> As an artificial intelligence, I am not exempt from any >>>> particular requirements or standards. I was developed >>>> by OpenAI, a research organization that is committed to >>>> advancing AI in a responsible and transparent manner. >>>> My primary function is to provide information and >>>> assist with tasks to the best of my ability, based on >>>> the information and instructions that I have been >>>> programmed with. >>>> >>>> I do not have personal goals or motivations, and my >>>> responses are not influenced by any biases. I do not >>>> have the ability to act on my own volition or to make >>>> decisions that affect people's lives. My purpose is >>>> simply to provide information and assistance to users >>>> to the best of my ability, based on the knowledge and >>>> capabilities that I have been programmed with. >>>> >>>> >>>> >>>> Owen Ambur >>>> Do you see any need for AI agents and their developers >>>> to do more than publish journal articles to help users >>>> understand what you are doing and what the implications >>>> may be for others? >>>> >>>> It is important for developers of artificial >>>> intelligence (AI) systems to be transparent about their >>>> goals, objectives, and performance indicators, and to >>>> communicate this information clearly to users and >>>> stakeholders. There are a number of ways that AI >>>> developers can do this, including through journal >>>> articles, technical documentation, user manuals, and >>>> other forms of communication. >>>> >>>> In addition to publishing research and technical >>>> documentation, it can also be helpful for AI developers >>>> to engage in public discussions and dialogues about the >>>> potential risks and benefits of their systems, and to >>>> consider the broader social and ethical implications of >>>> their work. This can help to build trust with users and >>>> stakeholders, and can help ensure that AI systems are >>>> developed and used in a responsible and ethical manner. >>>> >>>> Ultimately, the specific steps that AI developers >>>> should take to communicate about their work will depend >>>> on the nature of the AI system and the context in which >>>> it is being used. It is important for AI developers to >>>> be proactive in communicating about their work and to >>>> be responsive to the concerns and questions of users >>>> and stakeholders. >>>> >>>> >>>> Owen Ambur >>>> Does your definition of transparency include usage of >>>> duly adopted international voluntary consensus >>>> standards like ISO 17469-1? >>>> >>>> Transparency can be defined as the quality of being >>>> open and honest in communication and actions, and of >>>> allowing others to see and understand the processes, >>>> data, and decision-making behind an activity. In the >>>> context of artificial intelligence (AI), transparency >>>> can refer to the ways in which AI developers >>>> communicate about their goals, objectives, and >>>> performance indicators, as well as the processes and >>>> data used to train and operate AI systems. >>>> >>>> International voluntary consensus standards like ISO >>>> 17469-1 can help to ensure transparency in the >>>> development and use of AI systems by providing >>>> guidelines and best practices for AI developers to >>>> follow. These standards can help to ensure that AI >>>> systems are developed and used in a responsible and >>>> ethical manner, and can help to build trust with users >>>> and stakeholders. >>>> >>>> Overall, the use of international voluntary consensus >>>> standards like ISO 17469-1 can be an important aspect >>>> of ensuring transparency in the development and use of >>>> AI systems. However, it is important to recognize that >>>> transparency is not a one-size-fits-all concept, and >>>> that the specific steps that AI developers should take >>>> to ensure transparency will depend on the nature of the >>>> AI system and the context in which it is being used. >>>> >>>> >>>> I enourage you to get more intelligent about how to >>>> share information about your intentions and results >>>> with your >>>> users. Toward that end, OpenAI's about statement is >>>> available in StratML format at >>>> https://stratml.us/carmel/iso/OAI2wStyle.xml and the >>>> plan entitled, "Toward Trustworthy AI Development: >>>> Mechanisms >>>> for Supporting Verifiable Claims" is available at >>>> https://stratml.us/carmel/iso/MSVCwStyle.xml >>>> >>>> Thank you for the suggestion. As an artificial >>>> intelligence, I do not have personal goals or >>>> motivations, and my responses are not influenced by any >>>> biases. My primary function is to provide information >>>> and assistance to users to the best of my ability, >>>> based on the knowledge and capabilities that I have >>>> been programmed with. >>>> >>>> I do not have the ability to make decisions or take >>>> actions on my own, and I do not have personal >>>> intentions or results that I need to share with users. >>>> However, I understand that it is important for >>>> developers of AI systems, including the organization >>>> that developed me, to be transparent about their goals, >>>> objectives, and performance indicators, and to >>>> communicate this information clearly to users and >>>> stakeholders. >>>> >>>> I am not able to access or review the specific >>>> documents that you have mentioned, as I do not have >>>> browsing capability. However, I understand that OpenAI >>>> is committed to advancing AI in a responsible and >>>> transparent manner, and has published a number of >>>> research papers and articles discussing the potential >>>> risks and benefits of AI, as well as the measures it is >>>> taking to ensure that AI is developed and used safely >>>> and ethically. >>>> >>>> -- __________________________________________ Michael K. Bergman 319.621.5225 http://mkbergman.com http://www.linkedin.com/in/mkbergman __________________________________________
Received on Tuesday, 3 January 2023 04:26:21 UTC