- From: Paola Di Maio <paoladimaio10@gmail.com>
- Date: Wed, 11 Jan 2023 11:12:37 +0800
- To: Dave Raggett <dsr@w3.org>
- Cc: W3C AIKR CG <public-aikr@w3.org>
- Message-ID: <CAMXe=SrQZ6_bJP-njoTJ9VXcYXSH7ANde9Hwin9y5EC0fPFbeA@mail.gmail.com>
Thank you for the analysis On Tue, Jan 10, 2023 at 7:15 PM Dave Raggett <dsr@w3.org> wrote: > On 10 Jan 2023, at 04:56, Paola Di Maio <paola.dimaio@gmail.com> wrote: > > Not saying anything in favour or against this paper but > related to AI KR , hence worth a glance > > Constitutional AI: Harmlessness from AI Feedback > https://www.anthropic.com/constitutional.pdf > > > Thanks for the pointer. > > The work starts by selecting an initial prompt from a dataset of harmful > prompts. This is followed by prompts to request a self-critique, to request > a revision, and then repeating the original harmful prompt to see if the > response is an improvement on the original response. This process can be > repeated with variations in the wording for the critique request and > revision request. The results can later be used to refine the large > language model to produce less harmful responses without the need for the > critique request and revision request steps. > > This relies on the large language model having sufficient understanding > to generate its own critiques and revisions. However, large language > model tend to be weak on reasoning. In particular, the deep learning > architectures used in LLMs, don’t learn to emulate reasoning functions. > Instead, they find clever ways to learn statistical features that > inherently exist in the reasoning problems. > > See: > https://bdtechtalks.com/2022/06/27/large-language-models-logical-reasoning/ > > That isn’t surprising as deep learning struggles with compositional > generalisability, relying on vast datasets in compensation. > > See: https://royalsocietypublishing.org/doi/10.1098/rstb.2019.0307 > And more generally: > https://link.springer.com/article/10.1007/s13748-021-00239-1 > > Existing large language models support chain of thought reasoning, relying > on working memory (the current neural activation values) to hold the > context. There is no support for continuous learning and the equivalent of > human short and long term memory. This is an opportunity for experimenting > with new network architectures and training techniques that would reduce > the need for vast datasets and better reflect what we know about human > cognition. > > Dave Raggett <dsr@w3.org> > > > >
Received on Wednesday, 11 January 2023 03:16:38 UTC