- From: Paola Di Maio <paola.dimaio@gmail.com>
- Date: Sun, 7 Jul 2019 19:42:48 +0800
- To: Dave Raggett <dsr@w3.org>
- Cc: Owen Ambur <owen.ambur@verizon.net>, W3C AIKR CG <public-aikr@w3.org>
- Message-ID: <CAMXe=SrbY43yQ3Dq=ExzfidpAo-zaYso4-RKjU-t61YU7-gJ4w@mail.gmail.com>
Dave, actually interested but no even a sliver of availability. so please do keep us posted here on what you are doing. Owen and Dave thanks for the replies. please clarify the connection of this to my anxiety about disentanglement cheers PDM On Sun, Jul 7, 2019 at 3:30 AM Dave Raggett <dsr@w3.org> wrote: > Hmm, a simpler interpretation is that feelings and emotions are > computations that guide our behaviour in respect to our goals and our > social interactions with others. Some of this further relates to fast vs > slow modes of thinking as popularised by Daniel Kahneman: > > "System 1 and System 2 are two distinct modes of decision making: System > 1 is an automatic, fast and often unconscious way of thinking. It is > autonomous and efficient, requiring little energy or attention, but > is prone to biases and systematic errors. System 2 is an effortful, slow > and controlled way of thinking." > > This is all too evident in how people think about politics, and for me, > suggests that as we work on developing strong AI, we need to ensure that AI > systems have feelings along with empathy and compassion, and avoid the lazy > ways of thinking that far too many humans use in respect to politics and > society. > > If anyone is actually interested in working on the practical aspects of > this, please contact me directly. > > On 6 Jul 2019, at 18:50, Owen Ambur <owen.ambur@verizon.net> wrote: > > In *Incognito: The Secret Lives of the Brain* > <http://www.eagleman.com/incognito>, David Eagleman downplays the role of > consciousness in determining our behavior, most of which is on autopilot. > > https://www.linkedin.com/pulse/consciously-connected-communities-owen-ambur/ > > > In *Against Empathy: The Case for Rational Compassion*, Paul Bloom says, "When > some people think about empathy, they think about kindness. I think about > war." (p. 188) > > While the math eludes me, the broader logic seems clear: > > Do we want to use our powers of reasoning merely to justify our emotions, > after-the-fact, as seems to be natural for us? And should we use AI to > augment (accentuate) the expression of our emotions ... as "social" > networking services tend to do? (It seem like mind altering drugs might be > more efficiently and effectively applied for that purpose.) > > Or might we prefer to apply logic (math) to improve the outcomes of our > actions? > > > Which of those two alternatives might make us "feel" better (be more > satisfied) in the long run? > > > Dave Raggett <dsr@w3.org> http://www.w3.org/People/Raggett > W3C Data Activity Lead & W3C champion for the Web of things > > > > > > >
Received on Sunday, 7 July 2019 11:43:48 UTC