Re: AI KR , foundation models explained (talking about slippery things

Thanks Paola and Dave for bringing this issue up again. The argument
brought forward by Dave, which I also mentioned several times in the past,
that we humans ate taught three things (at least one may assume so) in
school. Information about subjects which may be everyday common themes or
knowledge in carefully crafted packages about subjects (domains of
discourse), packaged in lessons or lectures, methodologies and processes
for acquiring, analyzing and synthesizing data and turning these into
information and or knowledge, and at an academic or professional level
structuring the knowledge for internal use or for interdisciplinary use.

It seems with all the hype about AI and how to create AGI, we are
forgetting the importance of the latter two which are key to making sure we
make use of as little data as possible to be able to come up with
information or knowledge.

Any formal modeling should take this into consideration.

I read an interesting article about expander graphs (
https://www.quantamagazine.org/in-highly-connected-networks-theres-always-a-loop-20240607/)
which could loosely be used to model how we expand our knowledge and the
underlying structures.

An article in Scientific American draws attention to what seem an
inevitable future in the field of mathematics, i.e. AI being a "co-pilot"
for proving and structuring knowledge (
https://www.scientificamerican.com/article/ai-will-become-mathematicians-co-pilot/
).

Which again makes the case for formally capturing knowledge in expander
graphs which represent ontologies based structuring of knowledge.

I am sure I am stating nothing new, but the mathematics behind expander
graphs presents a novel approach to knowledge representation

Milton Ponson
Rainbow Warriors Core Foundation
CIAMSD Institute-ICT4D Program
+2977459312
PO Box 1154, Oranjestad
Aruba, Dutch Caribbean


On Sat, Jun 8, 2024 at 5:00 AM Dave Raggett <dsr@w3.org> wrote:

> Training foundation models for LLMs is kind of like getting them to learn
> about everything all at once, all mixed up. This  works thanks to the magic
> of gradient descent and back propagation, and addresses the challenge that
> understanding every day sentences requires a good grasp of common sense
> knowledge, creating a chicken and egg problem.
>
> Back propagation is very slow (look at typical values for the learning
> rate) and very different from how humans learn. We are able to learn from
> single examples, and get by on a very tiny fraction of the data that LLMs
> require for foundation models. Chomsky referred to this as the “poverty of
> the stimulus”.
>
> During childhood, our schooling introduces knowledge in a carefully
> organised approach with new knowledge layered on top of previously learnt
> knowledge.  Our grasp of common sense comes from a blend of everyday
> experience and what we are schooled.
>
> In the last ten years AI has come a long way, but we are still to figure
> out how to mimic the economies of human learning. I am searching for the
> means for neural networks to memorise and generalise from sequences using
> single-shot learning. This means stepping away from back propagation to
> consider other, more biologically plausible approaches.  One paper that
> caught my eye combines slow learning for learning to learn, and fast
> learning for single-shot learning. In essence, this trains the network to
> learn quickly for a limited set of tasks.
>
> Tomorrow’s AI will be very different from today’s as we gradually master
> quick learning and deliberative (Type 2) reasoning. Moreover, it will use a
> fraction of the power consumed by today’s energy hungry GPUs/TPUs.  Sparse
> spiking neural networks implemented with neuromorphic hardware will mimic
> the efficiency of the brain.  This is also likely to trigger a move away
> from back propagation.
>
> There is a lot to look forward to.
>
> Cheers,
>     Dave
>
> On 8 Jun 2024, at 06:03, Paola Di Maio <paola.dimaio@gmail.com> wrote:
>
> Okay, folks, I have been a bit AWOL, got lost in the dense forest of
> understanding following the AI KR path
> In related discussions, what are foundation models?
>
> If you ask Google (exercise)  the answer points to FM in ML, starting with
> Stanford in 2018 etc etc etc
> https://hai.stanford.edu/news/what-foundation-model-explainer-non-experts
> Great resources to be found online, all pointing to ML and nobody actually
> showing you the FM
> is in a tangible form (I remember this happened a lot with SW)
> Apparently
> that FM are actually not an actual thing, they are not there at all,
>  they are like dynamic neural network architecture (no wonder they have
> been slippery all along) which is built by ingesting
> data on the internet
>
> *Foundation models are massive neural network-based architectures designed
> to process and generate human-like text. They are pre-trained on a
> substantial corpus of text data from the internet, allowing them to learn
> the intricacies of language, grammar, context, and patterns.*
>
> They are made of layers, heads and parameters
>
>
> Coming from systems engineering, you know, with a bit of an existential
> background, I am making the case
> that foundational models without ontological basis are actually the cause
> of much risk in AI
>
> In case you people were wondering what I am up to, and would like to
> contribute to this work
> Please pitch in
>
> Paola
>
>
> Dave Raggett <dsr@w3.org>
>
>
>
>

Received on Sunday, 9 June 2024 17:24:11 UTC