- From: Timothy Holborn <timothy.holborn@gmail.com>
- Date: Fri, 28 Oct 2022 23:33:14 +1000
- To: ProjectParadigm-ICT-Program <metadataportals@yahoo.com>
- Cc: Dave Raggett <dsr@w3.org>, Paola Di Maio <paoladimaio10@gmail.com>, W3C AIKR CG <public-aikr@w3.org>
- Message-ID: <CAM1Sok1Emct5a8t_7nbAy6Pm3+2mj1kWAYsTcMP3VbqCip7QGw@mail.gmail.com>
FWIW: I think the idea of artificial minds being rendered consciousness is an "ungodly" concept. Artificial minds being rendered in relation to property rights laws / asset related considerations, entirely plausible. I therefore think it's too dangerous to try to support peoples extension of self (digital twins) as it's likely to be something companies want to "own". Whereas; the idea of democratisation ownership of AI agents - or robots (whether their in a phone or some sort of others physical object) doesn't really matter. https://twitter.com/WebCivics/status/1585976653867405312 If humanity is under attack by dangerous robots, I'd like to have one that I own fighting for me, kinda like r2d2 but different. Timh. On Fri, 28 Oct 2022, 11:25 pm ProjectParadigm-ICT-Program, < metadataportals@yahoo.com> wrote: > There may be a relatively easy way out of this confusion. But it starts > with disentangling knowledge representation completely from AI. > > Following Dave Raggett's line of reasoning we posit knowledge > representation to be a class of semiotic (input) structured descriptions > that lend themselves to analysis through logical, computational, > mathematical and computability processes in order for these to create > computable (output) algorithms, given a certain set of objects in an object > system in physical reality (spatiotemporal defined set of confined spaces > and objects therein) which together with a set of relevant interaction > processes defining an interaction system. > > This way we eliminate the problem of distinguishing between structured > data, information and knowledge. > > For this interaction system we now define classes of transformational > mappings for the interaction system, (1) dealing with sensory input through > observation, (2) converting the observation datasets to formats to compare > to existing instances in the structured descriptions, (3) exchanging or > passing observed datasets to another structured description, (4) add, > delete, edit or deprecate instances to the structured description, (5) > trigger actions in the interaction system.. > > We can now use all mathematical, computer science, computability, and > mathematical tools from theoretical physics, representation theory, and > category theory to produce generalizations of the basic components, being > structured descriptions and interaction systems to build increasingly > complex sets. > > Note that the concepts of mind, consciousness and sell awareness are > avoided, but openness and explainability become embedded. > > Mind and consciousness come into play if we contemplate artificial general > intelligence. > > And in doing so we avoid any ontological and epistemological discussions > with philosophers, because those only arise at the AGI level. > > Milton Ponson > GSM: +297 747 8280 > PO Box 1154, Oranjestad > Aruba, Dutch Caribbean > Project Paradigm: Bringing the ICT tools for sustainable development to > all stakeholders worldwide through collaborative research on applied > mathematics, advanced modeling, software and standards development > > > On Thursday, October 27, 2022 at 09:05:10 PM AST, Paola Di Maio < > paoladimaio10@gmail.com> wrote: > > > Thank you all for contributing to the discussion > > the topic is too vast - Dave I am not worried if we aree or not agree, the > universe is big enough > > To start with I am concerned whether we are talking about the same thing > altogether. The expression human level intelligence is often used to > describe tneural networks, but that is quite ridiculous comparison. If the > neural network is supposed to mimic human level intelligence, then we > should be able to ask; how many fingers do humans have? > But this machine is not designed to answer questions, nor to have this > level of knowledge about the human anatomy. A neural network is not AI in > that sense > it fetches some images and mixes them without any understanding of what > they are > and the process of what images it has used, why and what rationale was > followed for the mixing is not even described, its probabilistic. go figure. > > Hay, I am not trying to diminish the greatness of the creative neural > network, it is great work and it is great fun. But a) it si not an artist. > it does not create something from scratch b) it is not intelligent really, > honestly,. try to have a conversation with a nn > > This is what KR does: it helps us to understand what things are and how > they work > It also helps us to understand if something is passed for what it is not > *(evaluation) > This is is why even neural network require KR, because without it, we don > know what it is supposed > to do, why and how and whether it does what it is supposed to do > > they still have a role to play in some computation > > * DR Knowledge representation in neural networks is not transparent, * > *PDM I d say that either is lacking or is completely random* > > > DR Neural networks definitely capture knowledge as is evidenced by their > capabilities, so I would disagree with you there. > > > PDM capturing knowledge is not knowledge representation, in AI, > capturing knowledge is only one step, the categorization of knowledge is > necessary to the reasoning > > > > > > > *We are used to assessing human knowledge via examinations, and I don’t > see why we can’t adapt this to assessing artificial minds * > because assessments is very expensive, with varying degrees of > effectiveness, require skills and a process - may not be feasible when AI > is embedded to test it/evaluate it > > > We will develop the assessment framework as we evolve and depend upon AI > systems. For instance, we would want to test a vision system to see if it > can robustly perceive its target environment in a wide variety of > conditions. We aren’t there yet for the vision systems in self-driving cars! > > Where I think we agree is that a level of transparency of reasoning is > needed for systems that make decisions that we want to rely on. Cognitive > agents should be able to explain themselves in ways that make sense to > their users, for instance, a self-driving car braked suddenly when it > perceived a child to run out from behind a parked car. We are less > interested in the pixel processing involved, and more interested in whether > the perception is robust, i.e. the car can reliably distinguish a real > child from a piece of newspaper blowing across the road where the newspaper > is showing a picture of a child. > > It would be a huge mistake to deploy AI when the assessment framework > isn’t sufficiently mature. > > Best regards, > > Dave Raggett <dsr@w3.org> > > > >
Received on Friday, 28 October 2022 13:33:40 UTC