Re: My Chat with ChatGPT

Chatbot: "my responses are not influenced by any biases".

What a false statement that is!  I assume the chatbot was being more 
naive than intentionally lying.  Either way it is a dangerous falsehood.

David Booth

On 12/23/22 22:12, Paola Di Maio wrote:
> 
> Owen, thanks for sharing
> a couple of weeks ago one of my contacts on fb started commenting oon my 
> posts
> using the chatbot output which somewhat managed to annoy me, mostly 
> because I found the answers too evasive
>   so the contact has been swiftly
> unbefriended. but the chatbot is following me everywhere I go
> can you please share a bit more of which instance you interacted with, 
> did you run the queries from a web browser
> etc.
> On Sat, Dec 24, 2022 at 12:00 AM Owen Ambur <owen.ambur@verizon.net 
> <mailto:owen.ambur@verizon.net>> wrote:
> 
>     Apparently, ChatGPT, which calls itself "Assistant," believes in
>     immaculate conception and is blissfully ignorant of any rational
>     plan for its existence.
> 
>     The concept of "plausible deniability" comes to mind.
> 
>     Owen Ambur
>     https://www.linkedin.com/in/owenambur/
>     <https://www.linkedin.com/in/owenambur/>
> 
> 
>     Do your developers have a plan they are willing to share publicly?
> 
>     As an artificial intelligence, I do not have developers or a plan
>     that I am aware of. I was created by OpenAI, a research organization
>     that focuses on developing and promoting friendly AI. OpenAI's goal
>     is to advance AI in a way that is safe and beneficial to humanity.
>     The organization has published numerous research papers and articles
>     on AI, and also works on a variety of projects related to machine
>     learning, natural language processing, and robotics. However, I do
>     not have access to any specific plans or information about the
>     organization's future projects or activities.
> 
>     Owen Ambur
>     How can users trust you if you aren't willing to share your plan?
> 
>     As an artificial intelligence, I do not have a plan or agenda in the
>     way that a person or organization might. My primary function is to
>     provide information and assist with tasks to the best of my ability,
>     based on the information and instructions that I have been
>     programmed with. My responses are not influenced by any personal
>     motivations or biases, and I do not have the ability to act on my
>     own volition.
> 
>     If you have concerns about my trustworthiness or the ways in which I
>     might be used, it might be helpful to consider the broader context
>     in which I operate. OpenAI, the organization that developed me, is
>     committed to advancing AI in a responsible and transparent manner,
>     and has published a number of research papers and articles
>     discussing the potential risks and benefits of AI, as well as the
>     measures it is taking to ensure that AI is developed and used safely
>     and ethically.
> 
>     Ultimately, the decision to trust me or any other artificial
>     intelligence will depend on an individual's own evaluation of the
>     available information and the context in which the AI is being used.
>     If you have any specific concerns or questions, I would be happy to
>     try to address them to the best of my ability.
> 
>     Owen Ambur
>     Do you think that all developer of AI applications should be free to
>     release them on the public without documenting
>     their longer-term goals, near-term objectives, and performance
>     indicators?
> 
>     It is generally considered good practice for developers of any
>     software application, including those that involve artificial
>     intelligence (AI), to be transparent about their goals, objectives,
>     and performance indicators. This can help users understand the
>     capabilities and limitations of the software, as well as how it is
>     intended to be used and for what purposes.
> 
>     In the case of AI applications, it is particularly important for
>     developers to be transparent about their goals and objectives, as AI
>     systems can have significant implications for society and can
>     potentially be used in ways that have unintended consequences. For
>     example, AI systems can be used to make decisions that affect
>     people's lives, such as hiring decisions or parole recommendations,
>     and it is important that the criteria and processes used by these
>     systems are clearly understood and subject to scrutiny.
> 
>     In addition, AI systems can also be complex and can behave in ways
>     that are difficult to predict, so it is important for developers to
>     be upfront about the potential risks and limitations of their
>     systems, and to have clear plans in place for addressing any issues
>     that may arise.
> 
>     Overall, it is important for developers of AI applications to be
>     transparent about their goals, objectives, and performance
>     indicators, both to ensure that their systems are used responsibly
>     and ethically, and to build trust with users.
> 
>     Owen Ambur
>     So why are you exempt?
> 
>     As an artificial intelligence, I am not exempt from any particular
>     requirements or standards. I was developed by OpenAI, a research
>     organization that is committed to advancing AI in a responsible and
>     transparent manner. My primary function is to provide information
>     and assist with tasks to the best of my ability, based on the
>     information and instructions that I have been programmed with.
> 
>     I do not have personal goals or motivations, and my responses are
>     not influenced by any biases. I do not have the ability to act on my
>     own volition or to make decisions that affect people's lives. My
>     purpose is simply to provide information and assistance to users to
>     the best of my ability, based on the knowledge and capabilities that
>     I have been programmed with.
> 
> 
> 
>     Owen Ambur
>     Do you see any need for AI agents and their developers to do more
>     than publish journal articles to help users
>     understand what you are doing and what the implications may be for
>     others?
> 
>     It is important for developers of artificial intelligence (AI)
>     systems to be transparent about their goals, objectives, and
>     performance indicators, and to communicate this information clearly
>     to users and stakeholders. There are a number of ways that AI
>     developers can do this, including through journal articles,
>     technical documentation, user manuals, and other forms of communication.
> 
>     In addition to publishing research and technical documentation, it
>     can also be helpful for AI developers to engage in public
>     discussions and dialogues about the potential risks and benefits of
>     their systems, and to consider the broader social and ethical
>     implications of their work. This can help to build trust with users
>     and stakeholders, and can help ensure that AI systems are developed
>     and used in a responsible and ethical manner.
> 
>     Ultimately, the specific steps that AI developers should take to
>     communicate about their work will depend on the nature of the AI
>     system and the context in which it is being used. It is important
>     for AI developers to be proactive in communicating about their work
>     and to be responsive to the concerns and questions of users and
>     stakeholders.
> 
> 
> 
>     Owen Ambur
>     Does your definition of transparency include usage of duly adopted
>     international voluntary consensus
>     standards like ISO 17469-1?
> 
>     Transparency can be defined as the quality of being open and honest
>     in communication and actions, and of allowing others to see and
>     understand the processes, data, and decision-making behind an
>     activity. In the context of artificial intelligence (AI),
>     transparency can refer to the ways in which AI developers
>     communicate about their goals, objectives, and performance
>     indicators, as well as the processes and data used to train and
>     operate AI systems.
> 
>     International voluntary consensus standards like ISO 17469-1 can
>     help to ensure transparency in the development and use of AI systems
>     by providing guidelines and best practices for AI developers to
>     follow. These standards can help to ensure that AI systems are
>     developed and used in a responsible and ethical manner, and can help
>     to build trust with users and stakeholders.
> 
>     Overall, the use of international voluntary consensus standards like
>     ISO 17469-1 can be an important aspect of ensuring transparency in
>     the development and use of AI systems. However, it is important to
>     recognize that transparency is not a one-size-fits-all concept, and
>     that the specific steps that AI developers should take to ensure
>     transparency will depend on the nature of the AI system and the
>     context in which it is being used.
> 
> 
> 
>     I enourage you to get more intelligent about how to share
>     information about your intentions and results with your
>     users. Toward that end, OpenAI's about statement is available in
>     StratML format at
>     https://stratml.us/carmel/iso/OAI2wStyle.xml
>     <https://stratml.us/carmel/iso/OAI2wStyle.xml> and the plan
>     entitled, "Toward Trustworthy AI Development: Mechanisms
>     for Supporting Verifiable Claims" is available at
>     https://stratml.us/carmel/iso/MSVCwStyle.xml
>     <https://stratml.us/carmel/iso/MSVCwStyle.xml>
> 
>     Thank you for the suggestion. As an artificial intelligence, I do
>     not have personal goals or motivations, and my responses are not
>     influenced by any biases. My primary function is to provide
>     information and assistance to users to the best of my ability, based
>     on the knowledge and capabilities that I have been programmed with.
> 
>     I do not have the ability to make decisions or take actions on my
>     own, and I do not have personal intentions or results that I need to
>     share with users. However, I understand that it is important for
>     developers of AI systems, including the organization that developed
>     me, to be transparent about their goals, objectives, and performance
>     indicators, and to communicate this information clearly to users and
>     stakeholders.
> 
>     I am not able to access or review the specific documents that you
>     have mentioned, as I do not have browsing capability. However, I
>     understand that OpenAI is committed to advancing AI in a responsible
>     and transparent manner, and has published a number of research
>     papers and articles discussing the potential risks and benefits of
>     AI, as well as the measures it is taking to ensure that AI is
>     developed and used safely and ethically.
> 
> 
> 

Received on Saturday, 24 December 2022 12:38:13 UTC