Re: unprecendente, hanging onto to knowledge models before AI takes them down

"nail in the coffin for symbolic knowledge representation"

Of course that's already been demonstrated to be false. The industry is
desperate for so-called "neuro-symbolic" solutions. Those won't be easy
either. I'm not holding my breath.

On Sat, May 31, 2025, 5:45 PM Paola Di Maio <paoladimaio10@gmail.com> wrote:

> What to say, Owen and Kevin
>
> Yesterday  a famous phrase in the Hinton Turing Award speech kept rolling
> into my head
> 'the nail in the coffin for symbolic knowledge representation' heralding
> an age of non logic based machine learning
>
> I attach the url and transcript for those who may want to listen to it
> again, it was such a good lecture btw,
> Too bad it played down symbolic/logic  AI
> Time to go back to those talks,
> This is where the troubles started (if not earlier) and kind of feels like
> it was a long time ago but 2018 ts just yesterday really
>
> I am enjoying every bit of AI, and I am also startled by its limitations
> (abandon logic and see what you get)
>
> Mind out, poor and deficient reasoning is not just a prerogative of AI,
> Humans excel and make errors. flawed conclusion and fallibility in general
> It is when AI becomes pervasive and starts interfering with our systems
> deleting our emails, rewriting our
> browser history that is going to be scary, when innocent people use the
> LLM to learn and write about a topic and do not realise
> that what they hear is only part of the story, however well written up and
> fast
>
> Again this is also true of all knowledge sources, bias is not something
> new, it has been part of records in world history
> But AI is now part of the interface that filters reality, and that is why
> it can become scary
>
> I have also seen bias and poor reasoning in the initiatives aimed at
> mitigating AI risks
>
> As long as we are aware I guess,to maintain that level of awareness in a
> dynamic requires paying a lot of attention to what is going on
> and that can only be done by a well tuned human brain
>
> https://www.annualreviews.org/content/journals/10.1146/annurev-neuro-062111-150525
>
> I want to also make a note of the transcript of the LLM output
> I made a mistake in my prompt, that tried to retrieve the Turing Award
> lecture mentioned above and wrote 2019, and the LLM hang on to the mistake
> throughout its response instead of correcting it. I attach two transcripts
> for reference only
>
>
>
>
>
> On Sun, Jun 1, 2025 at 1:35 AM Kevin Spellman <kevinfrsa@icloud.com>
> wrote:
>
>> Universal AI and LLM design as a regulated government responsibility
>> would bring accountability, uniformity, standards and ethics. Social media
>> and the algorithms that violate our digital rights only come to light when
>> we stumble on to it. LLM’s are based on our data and we did not clearly
>> agree to this (or at least I didn’t). There is an opacity on how they work,
>> how and what they are connected to and more so the steps in place to
>> mitigate bias as an example. In a field that is growing in complexity and
>> revenue, there are fewer safeguards and people to support and enforce a
>> standard for public and private AI handling our data.
>>
>> Please pardon the brevity
>> Sent from my iPhone
>>
>> *Dr. Kevin J Spellman, FRSA, CMRS*
>>
>> On 31 May 2025, at 16:17, Owen Ambur <owen.ambur@verizon.net> wrote:
>>
>> 
>> Paola, while it might be taken as self-serving flattery or, at least,
>> knowing your customer, ChatGPT's conclusion about the second of your two
>> references makes sense to me:
>>
>> Bottom Line
>>
>> Steven J. Vaughan-Nichols is voicing a legitimate warning: *if we train
>> AIs on trash, they will produce trash.* But the current reality is not
>> that AI is collapsing—it’s that the ecosystem around it is fragile and
>> poorly governed. The way forward isn't to abandon AI but to become more *intentional
>> and structured* in how we curate knowledge, govern inputs, and manage
>> usage.
>>
>> That’s where standards like StratML, structured data, and truly
>> responsible AI design can help avert the kind of collapse the article warns
>> about.
>>
>> The details of its argument are available here
>> <https://chatgpt.com/share/683b1bb1-14c0-800b-9d9a-381ce0935ec8>.
>>
>> Owen Ambur
>> https://www.linkedin.com/in/owenambur/
>>
>>
>> On Saturday, May 31, 2025 at 12:10:11 AM EDT, Paola Di Maio <
>> paola.dimaio@gmail.com> wrote:
>>
>>
>> Good  day
>>
>> I hope everyone gets a change to smell the flowers at least once a day
>>
>> As predicted, we are  rapidly rolling into a new age of AI driven
>> everything and  knowledge is all we ve got to understand what is happening
>> and how
>>
>> The changes are already impacting our individual and collective lives and
>> behaviours etc
>> and we won't even know (scratching head)
>>
>> The best that we can do is hang onto our instruments of discernment, KR
>> being one of them
>>
>> Two articles below bring up important points
>>
>> *Gemini may summarize your emails even if you dont opt it for the feature*
>>
>> https://techcrunch.com/2025/05/30/gemini-will-now-automatically-summarize-your-long-emails-unless-you-opt-out/
>>
>> Honestly I do not know if this is true. It may even be illegal and if it
>> depends on the geographi loation could end up being very confusing
>> for those who travel around a lot. How will it work, if one day a person
>> reads an email from one country and another day from another?
>> if someone is a Google insider enough, should be investigated imho
>>
>> *AI Model Collapse*
>> https://www.theregister.com/2025/05/27/opinion_column_ai_model_collapse/
>> When the AI models collapse all we are going to have left is going to be
>> the robust knowledge structure in our brain/minds and in our libraries
>>
>>
>> *Brace, brace*
>>
>>
>>

Received on Sunday, 1 June 2025 01:24:57 UTC