- From: Paola Di Maio <paola.dimaio@gmail.com>
- Date: Sun, 25 Dec 2022 09:38:45 +0800
- To: Owen Ambur <owen.ambur@verizon.net>
- Cc: "public-aikr@w3.org" <public-aikr@w3.org>, David Booth <david@dbooth.org>
- Message-ID: <CAMXe=SpQVVTbyV=AkkXrSQ9C3b76AJOJ6wZ2xvjWasWhkE9hiQ@mail.gmail.com>
It is still remarkable, as long as users are aware of the weaknesses/flaws which is true of any book, any source - understanding the limitations of what is being asserted is essential critical thinking- But it should definitely quote the sources - Last year, I came across a wonderful tool, the first to do what chatgpt is doing now it was called wikimedia assistant, by a guy called vladimir https://twitter.com/vladblagoje with whom I briefly corresponded (i wanted to put my hands on that lovely natural language query interface that gave informed, easy to read responses) Soon after I learned that wikimedia assistant had been acquired, and kind of disappeared I suspect it powers chatgpt. The readiness of wikimedia assistant showed me the readiness level for natural language interfaces Merry xmas everyone On Sun, Dec 25, 2022 at 12:22 AM Owen Ambur <owen.ambur@verizon.net> wrote: > Paola, yes, I used my Google credentials and accessed ChatGPT in Chrome, > at https://chat.openai.com/chat > > David, yes, the denial of bias caught my attention as well. I consider it > to be an example of artificial ignorance > <https://www.linkedin.com/pulse/artificial-ignorance-owen-ambur/>. > > Copied immediately below is a continuation of my chat. Note that > "Assistant" previously denied having been created by developers. Now, > however, it simply disclaims any "personal knowledge of the plans or > activities of my developers" ... thereby ascribing to itself personality > as well as ignorance of its own motivations. > > ChatGPT obviously has a rather large mind, in the sense that consistency > is the hobgoblin of small minds > <https://en.wikipedia.org/wiki/Wikipedia:Emerson_and_Wilde_on_consistency>. > Since hypocrisy is feature of politics rather than a bug, it might make a > great politician. > > On the other hand, unlike politics, I believe services like ChatGPT can be > improved to be truly useful.. if only they live up to the ideals their > developers are espousing. The first step is for them to document their > plans in an open, standard, machine-readable format. > > Owen Ambur > https://www.linkedin.com/in/owenambur/ > > Sounds like an example of plausible deniability to me. > > As an artificial intelligence, I do not have personal motivations or > agendas, and my responses are not influenced by any biases. My primary > function is to provide information and assistance to users to the best of > my ability, based on the knowledge and capabilities that I have been > programmed with. > > Plausible deniability refers to the ability to deny responsibility or > knowledge of something in a way that is believable, even if it is not true. > As an AI, I do not have the ability to act on my own volition or to make > decisions that affect people's lives, and I do not have personal knowledge > or responsibility for any actions or outcomes. > > I understand that it is important for developers of AI systems, including > the organization that developed me, to be transparent about their goals, > objectives, and performance indicators, and to communicate this information > clearly to users and stakeholders. I hope this helps to clarify my role and > capabilities. If you have any further questions, please don't hesitate to > ask. > [image: Owen Ambur] > Spoken like a true bureaucrat. > > As an artificial intelligence, I do not have personal motivations or > agendas, and my responses are not influenced by any biases. My primary > function is to provide information and assistance to users to the best of > my ability, based on the knowledge and capabilities that I have been > programmed with. I do not have the ability to act on my own volition or to > make decisions that affect people's lives. > > It is not accurate to describe me as a bureaucrat, as I do not hold any > position of authority or responsibility, and I do not have the ability to > make decisions or take actions on my own. I exist solely to provide > information and assistance to users, and my responses are based on the > information and instructions that I have been programmed with. If you have > any further questions, please don't hesitate to ask. > [image: Owen Ambur] > Do your programmers have a plan they might be willing to share with you > and does it document the sources of information you use to generate your > responses? > > As an artificial intelligence, I do not have personal knowledge of the > plans or activities of my developers, as I do not have the ability to > access or browse information or communicate with them directly. My > responses are based on the information and instructions that I have been > programmed with, and I do not have the ability to access or modify this > information. > > I was developed by OpenAI, a research organization that focuses on > developing and promoting friendly AI. OpenAI's goal is to advance AI in a > way that is safe and beneficial to humanity. The organization has published > numerous research papers and articles on AI, and also works on a variety of > projects related to machine learning, natural language processing, and > robotics. However, I do not have access to any specific plans or > information about the organization's future projects or activities. > > I am able to generate responses to user queries by accessing the vast > amount of information that I have been trained on, which includes a wide > range of sources such as books, articles, and websites. However, I do not > have the ability to browse the internet or access new information that has > become available since my knowledge cutoff date of 2021. If you have any > further questions about my capabilities or the sources of information that > I use to generate responses, please don't hesitate to ask. > > > > On Saturday, December 24, 2022 at 07:38:36 AM EST, David Booth < > david@dbooth.org> wrote: > > > Chatbot: "my responses are not influenced by any biases". > > What a false statement that is! I assume the chatbot was being more > naive than intentionally lying. Either way it is a dangerous falsehood. > > David Booth > > On 12/23/22 22:12, Paola Di Maio wrote: > > > > Owen, thanks for sharing > > a couple of weeks ago one of my contacts on fb started commenting oon my > > posts > > using the chatbot output which somewhat managed to annoy me, mostly > > because I found the answers too evasive > > so the contact has been swiftly > > unbefriended. but the chatbot is following me everywhere I go > > can you please share a bit more of which instance you interacted with, > > did you run the queries from a web browser > > etc. > > On Sat, Dec 24, 2022 at 12:00 AM Owen Ambur <owen.ambur@verizon.net > > <mailto:owen.ambur@verizon.net>> wrote: > > > > Apparently, ChatGPT, which calls itself "Assistant," believes in > > immaculate conception and is blissfully ignorant of any rational > > plan for its existence. > > > > The concept of "plausible deniability" comes to mind. > > > > Owen Ambur > > https://www.linkedin.com/in/owenambur/ > > <https://www.linkedin.com/in/owenambur/> > > > > > > Do your developers have a plan they are willing to share publicly? > > > > As an artificial intelligence, I do not have developers or a plan > > that I am aware of. I was created by OpenAI, a research organization > > that focuses on developing and promoting friendly AI. OpenAI's goal > > is to advance AI in a way that is safe and beneficial to humanity. > > The organization has published numerous research papers and articles > > on AI, and also works on a variety of projects related to machine > > learning, natural language processing, and robotics. However, I do > > not have access to any specific plans or information about the > > organization's future projects or activities. > > > > Owen Ambur > > How can users trust you if you aren't willing to share your plan? > > > > As an artificial intelligence, I do not have a plan or agenda in the > > way that a person or organization might. My primary function is to > > provide information and assist with tasks to the best of my ability, > > based on the information and instructions that I have been > > programmed with. My responses are not influenced by any personal > > motivations or biases, and I do not have the ability to act on my > > own volition. > > > > If you have concerns about my trustworthiness or the ways in which I > > might be used, it might be helpful to consider the broader context > > in which I operate. OpenAI, the organization that developed me, is > > committed to advancing AI in a responsible and transparent manner, > > and has published a number of research papers and articles > > discussing the potential risks and benefits of AI, as well as the > > measures it is taking to ensure that AI is developed and used safely > > and ethically. > > > > Ultimately, the decision to trust me or any other artificial > > intelligence will depend on an individual's own evaluation of the > > available information and the context in which the AI is being used. > > If you have any specific concerns or questions, I would be happy to > > try to address them to the best of my ability. > > > > Owen Ambur > > Do you think that all developer of AI applications should be free to > > release them on the public without documenting > > their longer-term goals, near-term objectives, and performance > > indicators? > > > > It is generally considered good practice for developers of any > > software application, including those that involve artificial > > intelligence (AI), to be transparent about their goals, objectives, > > and performance indicators. This can help users understand the > > capabilities and limitations of the software, as well as how it is > > intended to be used and for what purposes. > > > > In the case of AI applications, it is particularly important for > > developers to be transparent about their goals and objectives, as AI > > systems can have significant implications for society and can > > potentially be used in ways that have unintended consequences. For > > example, AI systems can be used to make decisions that affect > > people's lives, such as hiring decisions or parole recommendations, > > and it is important that the criteria and processes used by these > > systems are clearly understood and subject to scrutiny. > > > > In addition, AI systems can also be complex and can behave in ways > > that are difficult to predict, so it is important for developers to > > be upfront about the potential risks and limitations of their > > systems, and to have clear plans in place for addressing any issues > > that may arise. > > > > Overall, it is important for developers of AI applications to be > > transparent about their goals, objectives, and performance > > indicators, both to ensure that their systems are used responsibly > > and ethically, and to build trust with users. > > > > Owen Ambur > > So why are you exempt? > > > > As an artificial intelligence, I am not exempt from any particular > > requirements or standards. I was developed by OpenAI, a research > > organization that is committed to advancing AI in a responsible and > > transparent manner. My primary function is to provide information > > and assist with tasks to the best of my ability, based on the > > information and instructions that I have been programmed with. > > > > I do not have personal goals or motivations, and my responses are > > not influenced by any biases. I do not have the ability to act on my > > own volition or to make decisions that affect people's lives. My > > purpose is simply to provide information and assistance to users to > > the best of my ability, based on the knowledge and capabilities that > > I have been programmed with. > > > > > > > > Owen Ambur > > Do you see any need for AI agents and their developers to do more > > than publish journal articles to help users > > understand what you are doing and what the implications may be for > > others? > > > > It is important for developers of artificial intelligence (AI) > > systems to be transparent about their goals, objectives, and > > performance indicators, and to communicate this information clearly > > to users and stakeholders. There are a number of ways that AI > > developers can do this, including through journal articles, > > technical documentation, user manuals, and other forms of > communication. > > > > In addition to publishing research and technical documentation, it > > can also be helpful for AI developers to engage in public > > discussions and dialogues about the potential risks and benefits of > > their systems, and to consider the broader social and ethical > > implications of their work. This can help to build trust with users > > and stakeholders, and can help ensure that AI systems are developed > > and used in a responsible and ethical manner. > > > > Ultimately, the specific steps that AI developers should take to > > communicate about their work will depend on the nature of the AI > > system and the context in which it is being used. It is important > > for AI developers to be proactive in communicating about their work > > and to be responsive to the concerns and questions of users and > > stakeholders. > > > > > > > > Owen Ambur > > Does your definition of transparency include usage of duly adopted > > international voluntary consensus > > standards like ISO 17469-1? > > > > Transparency can be defined as the quality of being open and honest > > in communication and actions, and of allowing others to see and > > understand the processes, data, and decision-making behind an > > activity. In the context of artificial intelligence (AI), > > transparency can refer to the ways in which AI developers > > communicate about their goals, objectives, and performance > > indicators, as well as the processes and data used to train and > > operate AI systems. > > > > International voluntary consensus standards like ISO 17469-1 can > > help to ensure transparency in the development and use of AI systems > > by providing guidelines and best practices for AI developers to > > follow. These standards can help to ensure that AI systems are > > developed and used in a responsible and ethical manner, and can help > > to build trust with users and stakeholders. > > > > Overall, the use of international voluntary consensus standards like > > ISO 17469-1 can be an important aspect of ensuring transparency in > > the development and use of AI systems. However, it is important to > > recognize that transparency is not a one-size-fits-all concept, and > > that the specific steps that AI developers should take to ensure > > transparency will depend on the nature of the AI system and the > > context in which it is being used. > > > > > > > > I enourage you to get more intelligent about how to share > > information about your intentions and results with your > > users. Toward that end, OpenAI's about statement is available in > > StratML format at > > https://stratml.us/carmel/iso/OAI2wStyle.xml > > <https://stratml.us/carmel/iso/OAI2wStyle.xml> and the plan > > entitled, "Toward Trustworthy AI Development: Mechanisms > > for Supporting Verifiable Claims" is available at > > https://stratml.us/carmel/iso/MSVCwStyle.xml > > <https://stratml.us/carmel/iso/MSVCwStyle.xml> > > > > Thank you for the suggestion. As an artificial intelligence, I do > > not have personal goals or motivations, and my responses are not > > influenced by any biases. My primary function is to provide > > information and assistance to users to the best of my ability, based > > on the knowledge and capabilities that I have been programmed with. > > > > I do not have the ability to make decisions or take actions on my > > own, and I do not have personal intentions or results that I need to > > share with users. However, I understand that it is important for > > developers of AI systems, including the organization that developed > > me, to be transparent about their goals, objectives, and performance > > indicators, and to communicate this information clearly to users and > > stakeholders. > > > > I am not able to access or review the specific documents that you > > have mentioned, as I do not have browsing capability. However, I > > understand that OpenAI is committed to advancing AI in a responsible > > and transparent manner, and has published a number of research > > papers and articles discussing the potential risks and benefits of > > AI, as well as the measures it is taking to ensure that AI is > > developed and used safely and ethically. > > > > > > > >
Received on Sunday, 25 December 2022 01:42:26 UTC