Re: My Chat with ChatGPT

Thank you Mike

I have signed up and asked a couple of questions, but I would have to check
the answers, do not have much time to play much with it now - ok, busy
playing with other things-

I am sure your evaluation fits to some extent, but I would need more
in-depth evaluation system-wide to understand the capability better, m,y
feel that the architecture is heterogeneous (works in some ways in some
parts and other ways in other parts)

As to whether it is intelligent or not, its a matter of definition
We cannot compare human intelligence to machine intelligence yet, however
a machine is intelligent some if it can do information processing

Not all intelligence can be automated, but some intelligent processes can,
the whole point of intelligent systems

So much to read on the subject, so little time to talk about it
OLD TOPIC (paywalled)  Intelligence as an information-processing concept
Earl Hunt  First published: November 1980
https://doi.org/10.1111/j.2044-8295.1980.tb01760.x
Citations: 331

and  Open Access preprint, more recent
Intelligence as information processing: brains, swarms, and computers
Carlos Gershenson
<https://arxiv.org/search/cs?searchtype=author&query=Gershenson%2C+C>

There is no agreed definit

https://arxiv.org/abs/2108.05349



On Tue, Jan 3, 2023 at 12:26 PM Mike Bergman <mike@mkbergman.com> wrote:

> Hi Paola,
>
> I think this is state-of-the-art for textual Q & A systems. To learn
> further, it is likely essential to sign up for a free account
> <https://chat.openai.com/auth/login> (for now) and begin to play. To
> narrow or expand queries, think of your questions to the system as Venn
> phrases in a natural language construction with the standard AND or OR
> Boolean operators. To exclude things, you can say natural language things
> like, "Besides these [A, B, or C conditions]' or you can specify the
> relations such as 'similar' or 'different' or add the action Ws (who, what,
> where, when, why), etc. I personally always ask the system in a polite and
> respectful way. But I suspect the system does not require it; one could be
> demanding or snarky if that is your preferred style.
>
> The system apparently learns (short term? earlier threads), but I find
> comparing the gaps in my expected answers to my questions to be most
> helpful in refining my next query.
>
> I find the system to be an assistant, one who finds more connections or
> relations or helps one remember prior known connections. It is not
> infallible and misses many connections. But notice I said 'one' there. That
> is because, if not intelligent, the system is a knowledgeable and
> interactive agent. I don't know as we need to get more cosmic than that.
> Knowledge nuggets, yes; new insights, no. Still, pretty helpful in the
> right contexts.
>
> Best, Mike
> On 1/2/2023 9:35 PM, Paola Di Maio wrote:
>
> so maybe it is a little bit intelligent and can do some basic checks (ie,
> parsing a link from a source as a form of basic verification)
>
> thanks a lot for sharing
>
> On Tue, Jan 3, 2023 at 11:32 AM Naval Sarda <nsarda@epicomm.net> wrote:
>
>> The internals of how verification will happen is not known.
>>
>> I suppose the first person to record correction for a topic is going in
>> directly if the person is citing external link about the correct
>> information. Looks like same concept on which wikipedia works.
>>
>> I did site external link stating that it is link from companies website
>> who have manufatured roomba.
>>
>> Naval
>> On 03/01/23 8:40 am, Paola Di Maio wrote:
>>
>> Thank you Naval
>> now I understand better
>>
>> so, you are telling us that
>> a) the machine is learning
>> b) it is responding to interaction
>>
>> If so, this is the beauty of this tool, we can teach it things
>> veeeery powerful
>>
>> so, what about if another user (or you) change the process again to
>> something else
>> entirely, does the machine have a way of verifying that what the user is
>> inputting is plausible
>> a really intelligent machine would have ways of verifying the input
>> depending on the type of input, the verification mechanisms would also
>> change
>>
>> On Tue, Jan 3, 2023 at 11:03 AM Naval Sarda <nsarda@epicomm.net> wrote:
>>
>>> I was just chatting with chat GPT 3 like humans without using any
>>> external tool. In the chat interaction, I typed that you gave me wrong
>>> information about roomba i3 reboot and as per companies website link below
>>> your instructions we incorrect.
>>>
>>> Then again asked, can you tell me what is the correct way to reboot
>>> roomba,
>>>
>>> Then the response came with new set of instructions, which was again
>>> incorrect.
>>>
>>> Then I typed in chat box that steps are still incorrect. Please note the
>>> correct steps as below and pasted the correct steps in the chat box.
>>> And again asked in chat box, now please tell me the correct process to
>>> reboot roomba i3 and it started giving correct steps.
>>>
>>> Naval
>>> On 03/01/23 7:24 am, Paola Di Maio wrote:
>>>
>>> Naval
>>> I thats really great to hear but please give more info=
>>> resetting roomba? uh? please explain
>>> did you take screenshots?
>>> can the error/correction be replicated?
>>> (roll back and repeat the process?)
>>> what patch did you apply and was it applied to the whole system/domain
>>> or only locally to you.in relation to a knowledge branch?
>>> I need to know
>>> thanks
>>>
>>> On Tue, Jan 3, 2023 at 3:33 AM Naval Sarda <nsarda@epicomm.net> wrote:
>>>
>>>> I was able to correct some information Chat GPT was sharing.
>>>>
>>>> I asked Chat GPT for instructions to reboot roomba i3. It gave wrongs
>>>> instructions. Instructions were for resetting roomba and not rebooting.
>>>>
>>>> Then I found correct information on google and iRobots' website. I
>>>> pasted the URL to chat gpt. It changed the instructions for rebooting but
>>>> it was still incorrect. Then I copied and pasted the correct instructions
>>>> from iRobot's website. Then it started giving correct instructions.
>>>>
>>>> Naval
>>>> On 02/01/23 9:10 pm, ProjectParadigm-ICT-Program wrote:
>>>>
>>>>
>>>> https://www.scientificamerican.com/article/chatgpt-explains-why-ais-like-chatgpt-should-be-regulated/
>>>>
>>>> Milton Ponson
>>>> GSM: +297 747 8280
>>>> PO Box 1154, Oranjestad
>>>> Aruba, Dutch Caribbean
>>>> Project Paradigm: Bringing the ICT tools for sustainable development
>>>> to all stakeholders worldwide through collaborative research on applied
>>>> mathematics, advanced modeling, software and standards development
>>>>
>>>>
>>>> On Friday, December 23, 2022 at 12:00:36 PM AST, Owen Ambur
>>>> <owen.ambur@verizon.net> <owen.ambur@verizon.net> wrote:
>>>>
>>>>
>>>> Apparently, ChatGPT, which calls itself "Assistant," believes in
>>>> immaculate conception and is blissfully ignorant of any rational plan for
>>>> its existence.
>>>>
>>>> The concept of "plausible deniability" comes to mind.
>>>>
>>>> Owen Ambur
>>>> https://www.linkedin.com/in/owenambur/
>>>>
>>>>
>>>> Do your developers have a plan they are willing to share publicly?
>>>>
>>>> As an artificial intelligence, I do not have developers or a plan that
>>>> I am aware of. I was created by OpenAI, a research organization that
>>>> focuses on developing and promoting friendly AI. OpenAI's goal is to
>>>> advance AI in a way that is safe and beneficial to humanity. The
>>>> organization has published numerous research papers and articles on AI, and
>>>> also works on a variety of projects related to machine learning, natural
>>>> language processing, and robotics. However, I do not have access to any
>>>> specific plans or information about the organization's future projects or
>>>> activities.
>>>> [image: Owen Ambur]
>>>> How can users trust you if you aren't willing to share your plan?
>>>>
>>>> As an artificial intelligence, I do not have a plan or agenda in the
>>>> way that a person or organization might. My primary function is to provide
>>>> information and assist with tasks to the best of my ability, based on the
>>>> information and instructions that I have been programmed with. My responses
>>>> are not influenced by any personal motivations or biases, and I do not have
>>>> the ability to act on my own volition.
>>>>
>>>> If you have concerns about my trustworthiness or the ways in which I
>>>> might be used, it might be helpful to consider the broader context in which
>>>> I operate. OpenAI, the organization that developed me, is committed to
>>>> advancing AI in a responsible and transparent manner, and has published a
>>>> number of research papers and articles discussing the potential risks and
>>>> benefits of AI, as well as the measures it is taking to ensure that AI is
>>>> developed and used safely and ethically.
>>>>
>>>> Ultimately, the decision to trust me or any other artificial
>>>> intelligence will depend on an individual's own evaluation of the available
>>>> information and the context in which the AI is being used. If you have any
>>>> specific concerns or questions, I would be happy to try to address them to
>>>> the best of my ability.
>>>> [image: Owen Ambur]
>>>> Do you think that all developer of AI applications should be free to
>>>> release them on the public without documenting
>>>> their longer-term goals, near-term objectives, and performance
>>>> indicators?
>>>>
>>>> It is generally considered good practice for developers of any software
>>>> application, including those that involve artificial intelligence (AI), to
>>>> be transparent about their goals, objectives, and performance indicators.
>>>> This can help users understand the capabilities and limitations of the
>>>> software, as well as how it is intended to be used and for what purposes.
>>>>
>>>> In the case of AI applications, it is particularly important for
>>>> developers to be transparent about their goals and objectives, as AI
>>>> systems can have significant implications for society and can potentially
>>>> be used in ways that have unintended consequences. For example, AI systems
>>>> can be used to make decisions that affect people's lives, such as hiring
>>>> decisions or parole recommendations, and it is important that the criteria
>>>> and processes used by these systems are clearly understood and subject to
>>>> scrutiny.
>>>>
>>>> In addition, AI systems can also be complex and can behave in ways that
>>>> are difficult to predict, so it is important for developers to be upfront
>>>> about the potential risks and limitations of their systems, and to have
>>>> clear plans in place for addressing any issues that may arise.
>>>>
>>>> Overall, it is important for developers of AI applications to be
>>>> transparent about their goals, objectives, and performance indicators, both
>>>> to ensure that their systems are used responsibly and ethically, and to
>>>> build trust with users.
>>>> [image: Owen Ambur]
>>>> So why are you exempt?
>>>>
>>>> As an artificial intelligence, I am not exempt from any particular
>>>> requirements or standards. I was developed by OpenAI, a research
>>>> organization that is committed to advancing AI in a responsible and
>>>> transparent manner. My primary function is to provide information and
>>>> assist with tasks to the best of my ability, based on the information and
>>>> instructions that I have been programmed with.
>>>>
>>>> I do not have personal goals or motivations, and my responses are not
>>>> influenced by any biases. I do not have the ability to act on my own
>>>> volition or to make decisions that affect people's lives. My purpose is
>>>> simply to provide information and assistance to users to the best of my
>>>> ability, based on the knowledge and capabilities that I have been
>>>> programmed with.
>>>>
>>>>
>>>> [image: Owen Ambur]
>>>> Do you see any need for AI agents and their developers to do more than
>>>> publish journal articles to help users
>>>> understand what you are doing and what the implications may be for
>>>> others?
>>>>
>>>> It is important for developers of artificial intelligence (AI) systems
>>>> to be transparent about their goals, objectives, and performance
>>>> indicators, and to communicate this information clearly to users and
>>>> stakeholders. There are a number of ways that AI developers can do this,
>>>> including through journal articles, technical documentation, user manuals,
>>>> and other forms of communication.
>>>>
>>>> In addition to publishing research and technical documentation, it can
>>>> also be helpful for AI developers to engage in public discussions and
>>>> dialogues about the potential risks and benefits of their systems, and to
>>>> consider the broader social and ethical implications of their work. This
>>>> can help to build trust with users and stakeholders, and can help ensure
>>>> that AI systems are developed and used in a responsible and ethical manner.
>>>>
>>>> Ultimately, the specific steps that AI developers should take to
>>>> communicate about their work will depend on the nature of the AI system and
>>>> the context in which it is being used. It is important for AI developers to
>>>> be proactive in communicating about their work and to be responsive to the
>>>> concerns and questions of users and stakeholders.
>>>>
>>>> [image: Owen Ambur]
>>>> Does your definition of transparency include usage of duly adopted
>>>> international voluntary consensus
>>>> standards like ISO 17469-1?
>>>>
>>>> Transparency can be defined as the quality of being open and honest in
>>>> communication and actions, and of allowing others to see and understand the
>>>> processes, data, and decision-making behind an activity. In the context of
>>>> artificial intelligence (AI), transparency can refer to the ways in which
>>>> AI developers communicate about their goals, objectives, and performance
>>>> indicators, as well as the processes and data used to train and operate AI
>>>> systems.
>>>>
>>>> International voluntary consensus standards like ISO 17469-1 can help
>>>> to ensure transparency in the development and use of AI systems by
>>>> providing guidelines and best practices for AI developers to follow. These
>>>> standards can help to ensure that AI systems are developed and used in a
>>>> responsible and ethical manner, and can help to build trust with users and
>>>> stakeholders.
>>>>
>>>> Overall, the use of international voluntary consensus standards like
>>>> ISO 17469-1 can be an important aspect of ensuring transparency in the
>>>> development and use of AI systems. However, it is important to recognize
>>>> that transparency is not a one-size-fits-all concept, and that the specific
>>>> steps that AI developers should take to ensure transparency will depend on
>>>> the nature of the AI system and the context in which it is being used.
>>>>
>>>> I enourage you to get more intelligent about how to share information
>>>> about your intentions and results with your
>>>> users. Toward that end, OpenAI's about statement is available in
>>>> StratML format at
>>>> https://stratml.us/carmel/iso/OAI2wStyle.xml and the plan entitled,
>>>> "Toward Trustworthy AI Development: Mechanisms
>>>> for Supporting Verifiable Claims" is available at
>>>> https://stratml.us/carmel/iso/MSVCwStyle.xml
>>>>
>>>> Thank you for the suggestion. As an artificial intelligence, I do not
>>>> have personal goals or motivations, and my responses are not influenced by
>>>> any biases. My primary function is to provide information and assistance to
>>>> users to the best of my ability, based on the knowledge and capabilities
>>>> that I have been programmed with.
>>>>
>>>> I do not have the ability to make decisions or take actions on my own,
>>>> and I do not have personal intentions or results that I need to share with
>>>> users. However, I understand that it is important for developers of AI
>>>> systems, including the organization that developed me, to be transparent
>>>> about their goals, objectives, and performance indicators, and to
>>>> communicate this information clearly to users and stakeholders.
>>>>
>>>> I am not able to access or review the specific documents that you have
>>>> mentioned, as I do not have browsing capability. However, I understand that
>>>> OpenAI is committed to advancing AI in a responsible and transparent
>>>> manner, and has published a number of research papers and articles
>>>> discussing the potential risks and benefits of AI, as well as the measures
>>>> it is taking to ensure that AI is developed and used safely and ethically.
>>>>
>>>> --
> __________________________________________
>
> Michael K. Bergman
> 319.621.5225http://mkbergman.comhttp://www.linkedin.com/in/mkbergman
> __________________________________________
>
>

Received on Tuesday, 3 January 2023 04:54:30 UTC