- From: Marco Neumann <marco.neumann@gmail.com>
- Date: Wed, 26 Jun 2019 16:45:54 +0100
- To: Patrick J Hayes <phayes@ihmc.us>
- Cc: Amirouche Boubekki <amirouche.boubekki@gmail.com>, Chris Harding <chris@lacibus.net>, Dave Raggett <dsr@w3.org>, Paola Di Maio <paoladimaio10@gmail.com>, ProjectParadigm-ICT-Program <metadataportals@yahoo.com>, semantic-web <semantic-web@w3.org>, xyzscy <1047571207@qq.com>
- Message-ID: <CABWJn4R3V0ocewBbN+74o7zGwOuFYvdwTNyiTOsssyUUXo_G_A@mail.gmail.com>
Pat, for completness sake, where do you place description logic (DL) in this context as a foundation for semantics on the web? On Wed, Jun 26, 2019 at 4:29 PM Patrick J Hayes <phayes@ihmc.us> wrote: > A quick remark: > > On Jun 26, 2019, at 8:03 AM, Dave Raggett <dsr@w3.org> wrote: > > I very much agree and have been arguing for a blend of symbolic and > statistical techniques using insights from decades of work in Cognitive > Psychology. Rational belief is about what can be justified given prior > knowledge and past experience. > > > So far in this thread we have been talking about knowledge representation > notations. You are here talking about mechanisms, not quite the same topic. > I entirely agree about the need to put together symbolic and statistical, > but I don’t see any reason why the use of the statistical would change the > nature or the semantics of the symbolic. (Do you?) > > This is not infallible, but nonetheless very useful in practice. It can > support higher order reasoning, something that is essential for modelling > human reasoning. > > > What kind of higher-order reasoning are you referring to here? The term > ‘higher-order’ has various meanings. If you simply mean that the logic can > mention, describe and quantify over properties and relationships as > first-class entities, then I would agree; but versions of FOL, even RDF, > can do that. > > Here is a test case for what I called ‘classical higher-order’ in an > earlier message. Do these facts: > > (P a) > (Q b) > > entail this higher-order statement: > > exists (X) (X a) & (X b) > > ? If not, then your logic is not what I would call higher-order. > The higher-order derivation mentions the property lambda (x)( (P x) or > (Q x) ) > > Pat > > > On 26 Jun 2019, at 14:54, Chris Harding <chris@lacibus.net> wrote: > > Formal logic is just one aspect of human reasoning (applied more or less > correctly, depending on the human in question). Human reasoning has other > aspects, giving it capabilities that formal logic does not have. For > example, it can handle inconsistencies. If the goal of AI is to approximate > human reasoning using computers, then its representational structures must > go beyond those of formal logic. > > Patrick J Hayes wrote: > > > > On Jun 25, 2019, at 6:06 PM, Amirouche Boubekki < > amirouche.boubekki@gmail.com> wrote: > > > > Le mar. 25 juin 2019 à 19:23, Patrick J Hayes <phayes@ihmc.us> a écrit : > >> >> >> On Jun 23, 2019, at 5:35 PM, ProjectParadigm-ICT-Program < >> metadataportals@yahoo.com> wrote: >> >> Again, let us look at the issue at hand. Artificial intelligence requires >> we represent knowledge in some format. All forms brought to the fore so far >> stick to a pretty simple way of representing knowledge. >> >> >> Most (all?) of the KR proposals put forward in AI or cognitive science >> work have been some subset of first-order predicate logic, using a variety >> of surface notations. There are some fairly deep results which suggest that >> any computably effective KR notation will not be /more/ expressive than FO >> logic. So FOL seems like a good ‘reference’ benchmark for KR expressivity. >> > > > > "Computably effective KR" > > That is one of the issue I try to address. > > > KR notation will not be /more/ expressive than FO logic > > Citation? > > > OK, this will take a little exposition. Notice up front that I said the > results /suggest/ something, not that they establish it beyond all doubt. > > The main result in question is called Lindstrom’s theorem. What it says, > technically, is that any logic (= a descriptive KR notation with a clear > semantics) which satisfies two conditions must be no stronger than FOL. The > two conditions are (1) compactness and (2) downward Lowenheim-Skolem (L-S). > OK, I won’t try to prove this here, but it is a theorem, OK? So bear with > me while I try to give an intuitive account of what these two conditions > mean, and why they are plausibly required for computational effectiveness.. > They can be intuitively summarized as the conditions that proofs can be > finitely wide and finitely deep. > > Compactness means that if something follows from a set of sentences, then > it must follow from a finite subset of them. Put simply, proofs have to be > finitely “wide”. This might seem kind of obvious, but there are quite > natural logics which don’t satisfy it. For example, suppose we had some > axioms for arithmetic which enabled one to prove that 0<1 and 1<2 and 2<3 > and… so on for every numeral N. Can you infer that x<x+1 for every number > x? Seems obvious, but an actual proof of this would have infinitely many > inputs. Compactness rules out things like this. Computationally this seems > extremely plausible, since we cannot get an infinite proof into any > physical memory. > > The (downward) L-S theorem is a bit harder to grok. It says that if a set > of sentence in the logic has any satisfying interpretation, then it has a > countable one. So if you can show that there isn't a countable one, then > you know there isn’t one at all. So what? Well, the key point here has to > do with how inference machinery operates. All inference systems can be seen > as ways of surveying all possible interpretations, looking for > counterexamples. You know that B follows from A when you can show that > there are no counterxamples, ie no interpretations which make A and (not B) > true. If your survey of interpretations is systematic and thorough, then > your logical inference machinery is correct. But any computational search > process can only generate finite structures. Now, /countably/ infinite > structures are fine, because counterexamples will be finite and hence will > be found eventually (this is based on a classical result called Koenig’s > lemma). So, in brief, the L-S theorem condition means that a finite search > through possible countable interpretations (which is the best that can be > done with finite machines) can be an effective complete search, In other > words, proofs that are finitely deep are enough, if the logic satisfies > this condition. So logics that don’t (such as classical /higher-order/ > predicate logic) are kind of ruled out as computationally plausible logics > anyway. > > OK, this is a very abbreviated summary of the reasoning, but the main > takeaway point is that these conditions, although maybe a bit > abstruse-seeming, really are very plausible conditions for any reasonable > KR notation which comes with reasoning machinery. And Lindstrom’s theorem > is, well, a theorem. > > Hope this helps. > > > > So FOL seems like a good ‘reference’ benchmark for KR > > What about things like Probabilist Logic Network (or Bayesian networks)? > > > I do not know for sure, but I would guess that a result similar to > Lindstrom’s would apply to logics with any kind of truthvalues, including > probabilities. My own, much more subjective view, is that probabilities are > simply the wrong model for KR. For just one observation, people are > absurdly poor at making probability estimates. But I won't try to justify > this view here :-) > > > By the way, OpenCog projects was very suspicious of my work when I cited > RDF. If you are interested I can create a document describing how their > database > called atom space works, so called, hypergraph database. > > And the those people are not alone. Other people told me RDF is deadend in > terms of > of (modern) KR for AI. > > > I might agree with that conclusion. For AI purposes, RDF is absurdly weak > and inexpressive. But AI is not what it is trying to do. > > Pat > > > But still, I am here :) > > > >> >> >> What we should be looking for is a generalized form in which objects can >> be linked. The graph is an obvious form. >> But we are focusing to much on the nuts and bolts level. >> >> Since it is the generally accepted intention to use AI in all walks of >> professional, commercial, personal and academic life, we should be looking >> at the various ways of representing knowledge. >> >> >> Otherwise we end up creating knowledge representation silos. >> >> >> Avoiding KR silos was one of the primary goals of the entire semantic-web >> linked-data initiative. But this has many aspects. First, we need to agree >> to all use a common basic notation. Triples (=RDF =Knowledge Graph >> =JSON-LD) has emerged as the popular choice. Getting just this much >> agreement has taken 15 years and thousands of man-hours of strenuous effort >> and bitterly contested compromises, so let us not try to undo any of that, >> no matter what the imperfections are of the final choice. >> > > For the record, I don't try to undo that. As a new actor, I am working > toward it. As any newbie, I may ask some questions badly, that could lead > you to think that I want a revolution. > > >> The next stage, which we are just getting started on, involves agreeing >> on a common vocabulary for referring to things, or perhaps a universal >> mechanism for clearly indicating that your name for something means the >> same as my name for that same thing. This seems to be much harder than the >> semantic KR pioneers anticipated. >> > > Good question. > > >> The third stage involves having a global agreement on the ontological >> foundations of our descriptions, what used to be called the ‘upper level >> ontology’. This is where we get into actual metaphysical disagreements >> about the nature of reality (are physical objects extended in time? How do >> we handle vague boundaries? What are the relationships between written >> tokens, images, symbols, conventions and the things they represent? What is >> a ‘background’? What is a ‘shape’? Is a bronze statue the same kind of >> thing as a piece of bronze? What changes when someone signs a contract? >> Etc. etc., etc.) This is where AI-KR and more recently, applied ontology >> engineering (not to mention philosophy) has been working for the past 40 or >> 50 years, and I see very little hope of any clear agreements acceptable to >> a large percentage of the world’s users. >> > > Pragmatic self: forget about that part from specification? > > >> Category theory diagrams, graphs and Feynman diagrams are three well >> known forms of representing knowledge graphs, but only in semantic web >> technologies we specify tuples, a restrictive form of representation. >> >> Category diagrams and Feynman diagrams are meaningful only within highly >> restricted and formal fields (category theory and quantum physics, >> respectively) so have little to do with general KR. If your point is that >> diagrams are useful, one can of course point to many examples of them being >> useful to human users, but this does not make them obviously useful in >> computer applications. >> >> Tuples are not more restrictive than graphs, since a collection of tuples >> is simply one way to implement a graph. Tuple stores ARE graphs. >> > > I would not say: "tuple stores are just [property] graph". Because my > implementation is much different. But I agree tuple store are some kind of > graph. > > For the record, the idea of the n-tuple store (or chunks store) came from > the need to version a quad store to factor some code. > Later I discovered it could me useful in other contexts: provenance, > quality, space, some kind of time. > Again, the nstore, is a performance trick. What you can do with a triple > store you can do with nstore, > performance will be different, nstore should be faster. I am by no means > trying to force the WG to adopt the proposal I made on github > <https://github.com/w3c/sparql-12/issues/98>, > I hope to learn something from the conversation, and I already did. > > > >> Best wishes >> >> Pat Hayes >> >> >> Milton Ponson >> GSM: +297 747 8280 >> PO Box 1154, Oranjestad >> Aruba, Dutch Caribbean >> Project Paradigm: Bringing the ICT tools for sustainable development to >> all stakeholders worldwide through collaborative research on applied >> mathematics, advanced modeling, software and standards development >> >> >> On Sunday, June 23, 2019, 3:57:01 AM ADT, Paola Di Maio < >> paoladimaio10@gmail.com> wrote: >> >> >> >> >> Chunks are also used in NLP (which is part of/related to CS either way) >> aka tokens >> Various useful references come up on searching chunks as tokens >> >> https://docs.oasis-open.org/dita/v1.2/os/spec/archSpec/chunking.html >> >> https://www.oxygenxml.com/doc/versions/21.1/ug-editor/topics/eppo-chunking.html >> >> On Sun, Jun 23, 2019 at 1:12 AM Dave Raggett <dsr@w3.org> wrote: >> >> >> >> On 22 Jun 2019, at 14:54, Amirouche Boubekki < >> amirouche.boubekki@gmail.com> wrote: >> >> Le ven. 21 juin 2019 à 16:27, Dave Raggett <dsr@w3.org> a écrit : >> >> Researchers in Cognitive Science have used graphs of chunks to represent >> declarative knowledge for decades, and chunk is their name for an n-tuple. >> >> >> I tried to lookup "graph of chunks" related to cognitive science. I could >> not find anything interesting outside this white paper about "accelerating >> science" [0] that intersect with my goals. >> >> [0] >> https://cra.org/ccc/wp-content/uploads/sites/2/2016/02/Accelerating-Science-Whitepaper-CCC-Final2.pdf >> >> >> Chunks are used on cognitive architectures, such as ACT-R, SOAR and >> CHREST, and is inspired by studies of human memory recall, starting with >> George Miller in 1956, and taken further by a succession of researchers. >> Gobet et al. define a chunk as “a collection of elements having strong >> associations with one another, but weak associations with elements within >> other chunks.” Cognitive Science uses computational models as the basis for >> making quantitive descriptions of different aspects of cognition including >> memory and reasoning. There are similarities to Frames and Property Graphs. >> >> Dave Raggett <dsr@w3.org> http://www.w3.org/People/Raggett >> W3C Data Activity Lead & W3C champion for the Web of things >> >> >> >> >> >> >> >> > > -- > Regards > > Chris > ++++ > > Chief Executive, Lacibus <https://lacibus.net/> Ltd > chris@lacibus.net > > > Dave Raggett <dsr@w3.org> http://www.w3.org/People/Raggett > W3C Data Activity Lead & W3C champion for the Web of things > > > > > > > > -- --- Marco Neumann KONA
Received on Wednesday, 26 June 2019 15:47:14 UTC