- From: Milton Ponson <rwiciamsd@gmail.com>
- Date: Thu, 8 Aug 2024 12:18:45 -0400
- To: Dave Raggett <dsr@w3.org>
- Cc: W3C AIKR CG <public-aikr@w3.org>, public-cogai <public-cogai@w3.org>
- Message-ID: <CA+L6P4zbuq5Tr1MOCwt33yW3VZdZ86bU89QMhywcW7YvUbuh9A@mail.gmail.com>
The market already seems to have made up its mind. https://finance.yahoo.com/news/generative-ai-getting-kicked-off-191657116.html The two trillion dollar question is: where do we go from here? KRR for AI and unraveling the brain processes key to memory formation, storage, recall and cognition seem the way to go. The latter are subject to issues of causality, nonlocality, quantum effects and the free energy principle. Buddhism sheds some light on causality and quantum effects. But we have a long way to go in terms of figuring out the inner workings of the human brain and the mathematical modeling of such and the hard problem of consciousness and current clashes between computer scientists, mathematicians and philosophers make it very clear that we need to figure out the right questions as Dave mentioned. That may very well be one of the hardest nuts to crack. In mathematics we have the Langlands program. We may have to come up with something similar for defining and discovering the linkages and interactions between different theories and disciplines in cognitive science, computational biology, neuroscience, philosophy, AI and mathematical modeling (the latter including theoretical physics as related to quantum physics and quantum biology). This program will elucidate the right questions and consequently show the path to KRR. Milton Ponson Rainbow Warriors Core Foundation CIAMSD Institute-ICT4D Program +2977459312 PO Box 1154, Oranjestad Aruba, Dutch Caribbean On Thu, Aug 8, 2024 at 6:29 AM Dave Raggett <dsr@w3.org> wrote: > LLMs are designed to make statistical predictions for text continuations, > and I am amazed by how well they do that. It is unsurprising that they are > weak on semantic consistency, just as they are for learning from limited > data. LLMs are good for summarisation, but not for deep insights. That’s > fine for some applications, but not others. LLMs can certainly help with > web search, but you still need to think for yourself given the limitations > of LLMs. I very much doubt that today’s LLMs will provide the hoped for > return on investment. That said, LLMs are useful and will remain > commonplace. > > I am not sure that Buddhism can help much when it comes to the details of > how neurons compute. Next generation AI will depend on a much better > understanding of human memory. How is it that we can remember and learn > from single episodes? That includes speech and music. One of the > challenges is how neurons can learn on the scale of seconds and much longer > rather than milliseconds. I am trying to figure out which questions are > vital to developing a model that I can implement in code. Progress is > dependent on identifying the “right” questions. > > > On 7 Aug 2024, at 17:24, Milton Ponson <rwiciamsd@gmail.com> wrote: > > https://www.theguardian.com/technology/article/2024/aug/06/ai-llms. > > Interesting article that stresses the point that relationships between > facts differentiate humans (for now) and AI that uses stochastic > information about tokens to come up with text generation and problem > solving. > > Now here is the mind boggling aspect, Buddhists talk about dependent > arising in knowledge, the free energy principle, causal cognition and the > way the brain processes sensory input and assimilates and stores this all > hint at complex sequential processing across multiple areas of the brain > with particular wave activities surging back and forth between areas in the > brain. > > This seems to make LLM generative AI not fit for modeling AGI. > > > Milton Ponson > Rainbow Warriors Core Foundation > CIAMSD Institute-ICT4D Program > +2977459312 > PO Box 1154, Oranjestad > Aruba, Dutch Caribbean > > > Dave Raggett <dsr@w3.org> > > > >
Received on Thursday, 8 August 2024 16:19:05 UTC