- From: <hans.teijgeler@quicknet.nl>
- Date: Mon, 29 Jul 2019 13:15:19 +0200
- To: "'Stephen D. Williams'" <sdw@lig.net>, <paoladimaio10@googlemail.com>
- Cc: "'SW-forum'" <semantic-web@w3.org>
- Message-ID: <000a01d545fe$e8d87fd0$ba897f70$@quicknet.nl>
Stephen, Isn’t it so that, no matter how elementary you go, you always end up with an explanation in a natural language of those information elements? Regards, Hans 15926.org <http://data.15926.org/> ______________________________________________-- From: Stephen D. Williams <sdw@lig.net> Sent: maandag 29 juli 2019 12:04 To: paoladimaio10@googlemail.com Cc: SW-forum <semantic-web@w3.org> Subject: Re: Hard vs. Soft or Hard + Soft KR, Re: neural networks being purported as KR? Paola, As far as I've been able to determine, there is no universal definition of concept so I find the concept of non-concept to be weakly defined at best. I think 'concept' has a granularity of meaning relative to a particular context, usually more than a triple and less than a large knowledge graph. I intuitively and practically define a concept as a set of elementary statements, facts, or reactions that can be used for prediction or action in a usually cohesive way somewhat separate from other cohesive facts. Good examples would be each mathematical operator, rule, or complete technique. The actual size, in whatever measure, is secondary to the cohesiveness and usual granularity of use. Somewhat orthogonal to this, my research a few years ago led me to an area that doesn't seem to have been explored: The morphology of concepts, or rather the morphology of capturing most or all of the essence of concepts: There is a somewhat finite set of morphologies that can capture the essence of most concepts. This is distinctly a different idea than the seeming (pre-ML boom) universality of triple graphs. Paola, I think you can see a set of triples or logic clauses as being able to represent a concept but are having trouble seeing an equivalence with a trained connectionist network. While, without visualization or similar extensions, it may be difficult to identify the cohesive locality of concepts in a ML network, I don't have so much difficulty in imagining that they are there, interlocking and perhaps haphazardly related. They can still be active in just the same ways, just only indirectly visible. (There are some good examples of building in visualization that sometimes largely indicates what is going on.) I think examples could be useful. Imagine we are trying to capture the logic of traffic and walk lights, sensors, and weekly schedules in a system. We could represent all of the states, events, and rules in first order logic, some equivalent set of triples, direct coding in software. Or we could hand design a logic circuit, timer chips (the 555 TTL etc.), relays, flipflops, etc. Or instead of designing such a circuit, we could assemble more than enough sensors, logic circuits, connection wires, signal lights, etc., just trying every combination in brute force but with no logic or intelligence until we produced a circuit that embodied the logic needed. Or, we could connect all of the sensors, switches, and add some test signals that indicate correct operation (also needed for the brute force solution and useful for testing the logic-based systems) and allow a ML system to converge on a working system through training on the error signal. Both the brute force circuit search and the ML result embody the same logic, the concepts in the explicitly logical approaches, even if they seem somewhat opaque. In a more complex situation, requiring internal or intermediate results that apply in different ways to many other cases, finding a solution with brute force may take a very, very long time to find. ML methods have been evolving to handle more depth and nuance in various ways. Stephen On 7/29/19 1:52 AM, Paola Di Maio wrote: Stephen thank you, I am thinking along those lines and we could have a discussion about hard vs soft KR My challenge today is to find a reference for ''non concept' (as used here Concept Recognition with Convolutional Neural Networks to Optimize Keyphrase Extraction Andreas Waldis , Luca Mazzola(B) , and Michael Kaufmann School of Information Technology, Lucerne University of Applied Sciences, 6343 Rotkreuz) The notion of non concept seems to be increasingly used in ANN to contrast 'the notion of concept in traditional KR, But I cannot find out if this has been formally defined *I doubt it? So no only a lot of peer reviewed scholarly articles make false statements (say things which are not true) but also reference notions which have never been formally defined (most of this vague research seems to be happening in Switzerland) This is not science, is it? Together with other things I am not figuring out, I tind this unsettling and adding to uncertainties PDM On Mon, Jul 29, 2019 at 11:51 AM Stephen D. Williams <sdw@lig.net <mailto:sdw@lig.net> > wrote: Thanks, great info! Knowledge graphs, semantic web, and other explicit knowledge representations are important. But, as you point out, they currently don't capture everything, especially not probability and complex rational but not simple logic relationships. ML and related enabled training, capture, and usage of this aspect of information, but most or all structure is opaquely inferred and usually shallow. Somehow we need all of the capabilities of both in a merged or married structure. I totally believe in the value of both. My observation is that we almost certainly reason with structure, logic, math in general, language, etc. on top of ML-like mechanisms (albeit far more varied, numerous, and expansive), and we fill in the gaps all the time with ML-like sub-rational or at least sub-logical reasoning. Most of us most of the time reason largely in this sub-logical reasoning mode, often interpreting, viewing, or fitting those thoughts and conclusions to formal structures afterwards or in a trailing overlapping fashion. https://www.ted.com/playlists/384/how_your_brain_constructs_real As a software & security architect and now robotic hardware designer, I never think in formal equations or systems, yet I automatically explore creative solution spaces with many, many constraints and patterns in mind. I can often construct equations of various types to summarize and communicate. I now find this to be true of business and interpersonal situations (various concerns, constraints, interests, conflicting aspects) and mechanical systems. For instance, this is a summary of part of the robotic actuator I've been working on: I1 * r/R * CVT - I2 * R/2 = RIVT I never could have found that solution through explicit logical deduction, yet the process of working through and mentally testing possibilities is very concrete and not mysterious at all. How do you experience your knowledge, thinking, problem solving, and creativity? Do you feel strongly and deeply aware of it or is it mysterious and opaque? Do you visualize a little, a lot, or not at all? (We're recently realized that something like 10% of people cannot visualize anything, ever, yet a number of animators are among those! They seem to use drawing as external working memory / thought space.) We already have ML solutions that are able to learn far more than simple table lookups. Creating a program to play a video game using strictly explicit logic, KG methods would be challenging while we have straight foward ML solutions already: https://thenextweb.com/artificial-intelligence/2018/08/23/researchers-gave-ai-curiosity-and-it-played-video-games-all-day/ https://towardsdatascience.com/an-exploration-of-neural-networks-playing-video-games-3910dcee8e4a It seems promising to extrapolate from these for some distance. Our challenge is to find a intersection method that allows us to combine these two mostly independent paths to create really interesting solutions. My sense is that there are two possibilities: We get an interface method working that allows an ML system to access computational algorithms / KG / semantic / logical knowledge through a kind of port, portal, map, or similar. Or that we evolve the ML systems to the point where they can embody, in a reasonably efficient and accurate way, this computational / KG / semantic / logical knowledge along with the more probabilistic, analog networked knowledge. Or both. Rather than connectionist vs. reductionist etc., perhaps these are useful labels: Hard knowledge: computational algorithms / KG / semantic / logical knowledge Soft Knowledge: ML, vision, everything connectionist or probabilistic Or sharp vs. round or curved or smooth. Stephen On 7/26/19 9:02 AM, ProjectParadigm-ICT-Program wrote: Thank you Paola for pointing this out. Again I must beat the drum. Knowledge is much more than extracting structure from facts and data. If I just recall that the collection of facts is subject to the uncertainty principle, any structure deduced cannot be complete, and the application of free will, and/or axiom of choice create a dichotomy, knowledge is much more. We are limited by our sensory apparatus, our hard wiring in our human brain, including the shortcuts made when processing visual data, and the limitations of natural language. I agree that knowledge reasoning should be fairly straightforward, but making the jump from KR to knowledge itself implies we come up with some consistent many worlds modeling scheme in which the virtual, mathematical and (many interpretations of) the physical world coexist, reconciling incompleteness, uncertainty principle, sensory limitations and application of free will and choice. A convergence of efforts by string theorists, researchers in human brain cognitive and biological structure fields, theoretical physicists and mathematicians working on finite groups, category theory, algebraic topology and logical structures for consistent super theories, and an odd mix of linguists and philosophers (including Buddhists) is doing just that. But they are far from a consensus. The point I am trying to make is that KR is more than semantics and ontologies and knowledge graphs, graphs, category theory diagrams and Feynmann diagrams and any other visualization tools we use. The implicate order David Bohm theoreticized underlying quantum reality and the reality of our physical world, cannot be captured by some mix of formal logic, semantic structures, ontologies or computable frameworks. And we we want someday A(G)I to be able to grasp human knowledge in general, we must create a growth path towards formal structures which have meta-layers above (knowledge) graphs, formal logic and ontologies. Mathematically speaking, using formal logic, ontologies and generalized graphs is necessary but insufficient for this general formal structure. And now I must add that deep learning and machine learning also fall short in terms of KR' If we let computer scientists, logicians, mathematicians and software engineers try to come up with KR which is fit for the AI we envision we will need for future applications we will fail miserably. We need neuroscientist, and specialists in the field of cognitive sciences, biologists and even psychologists, and philosophers and physicists to help us complete the general framework for knowledge, and to establish which parts can be effectively captured in a formal fashion, which provide suitable technologies and tools for KR. Mike Bergman did a nice expose on knowledge graphs at: A Common Sense View of Knowledge Graphs <http://www.mkbergman.com/2244/a-common-sense-view-of-knowledge-graphs/> A Common Sense View of Knowledge Graphs This article, based on a comprehensive history and definitions of the concept, provides a common-sense view of h... But historically even mandalas qualify as knowledge graphs, in a very stylized way. And they can be used to visualize very complex mathematical structures without the use of edges or arrows, thus removing the time component associated with the transition the edge or arrow represents, making knowledge representation in a time-independent fashion possible. Milton Ponson GSM: +297 747 8280 PO Box 1154, Oranjestad Aruba, Dutch Caribbean Project Paradigm: Bringing the ICT tools for sustainable development to all stakeholders worldwide through collaborative research on applied mathematics, advanced modeling, software and standards development On Thursday, July 25, 2019, 11:52:23 PM ADT, Paola Di Maio <mailto:paola.dimaio@gmail.com> <paola.dimaio@gmail.com> wrote: Sorry to bang on this topic, but its the task at hand at the moment I just found an article, which is good scientific survey then purports NN as a type of KR (casually sneaks in NN as the latest KR) This is published in a Springer peer reviewed publication and my makes all of my hairs stand up on my head This is the kind of rubbish that without further qualification is being passed down as the latest research, and which the future generations of AI scientists are being fed- wonder if anyone else has a problem with this proposition (sign of the times?) I am doing my best within my means to identify and contain this peril Article https://link-springer-com.nls.idm.oclc.org/article/10.1007/s00170-018-2433-8 A survey of knowledge representation methods and applications in machining process planning The machining process is the act of preparing the detailed operating instructions for changing an engineering design into an end product, which involves the removal of material from the part. Today, machining ... Xiuling Li, Shusheng Zhang, Rui Huang… in The International Journal of Advanced Manu… (2018) -- Stephen D. Williams Founder, Yebo, VolksDroid, Blue Scholar <tel:650-450-8649> 650-450-8649 | fax:703-995-0407 | <mailto:sdw@lig.net> sdw@lg.net | <https://HelloYebo.com> https://HelloYebo.com | <https://VolksDroid.org> https://VolksDroid.org | <https://BlueScholar.com> https://BlueScholar.com | https://sdw.st/in -- Stephen D. Williams Founder, Yebo, VolksDroid, Blue Scholar <tel:650-450-8649> 650-450-8649 | <fax:> fax:703-995-0407 | <mailto:sdw@lig.net> sdw@lg.net | <https://HelloYebo.com> https://HelloYebo.com | <https://VolksDroid.org> https://VolksDroid.org | <https://BlueScholar.com> https://BlueScholar.com | https://sdw.st/in --- Deze e-mail is gecontroleerd op virussen door AVG. http://www.avg.com
Received on Monday, 29 July 2019 11:15:49 UTC