In a recent
blog, I raised the question whether AI systems can have consciousness. Now
I found on the Psychology Today
website several articles by Marlynn
Wei that shed an interesting light on the question. In these articles
Marlynn Wei discusses recent AI research with psychological relevance.
In my blog I stated that having consciousness is not only a matter of showing consciousness-related
behaviour, but that it involves also having consciousness-related subjective
experiences. So, a conscious AI system should not only behave as if it is conscious,
but it should also have the right feelings, like having the right emotions in
the right situation. Having emotions is a complicated affair, but having the
right emotions in the right situation at least involves being able to recognize
the emotions of other humans, being able to have the right emotions in reaction
to the emotions others have, and reacting in the right way to emotions others
have. These three aspects of having emotions are not independent of each other,
as the discovery of the so-called mirror neurons has made
clear. If one of these three aspects of having emotions is missing, then we can
say that an AI system doesn’t have consciousness in my sense.
Although such conscious AI systems are still far away and don’t (yet?) exist, some
research discussed by Dr Wei is very interesting in this respect. For although,
for example, ChatGPT has gained
widespread attention for its ability to perform natural language processing
tasks, its skills go much farther than “only” producing texts. This chatbot is
also able to recognize and describe emotions. Moreover, it does it better than
humans do. At least this was the outcome of a recent
study by Zohar Elyoseph and others. Using a test called the Levels of
Emotional Awareness Scale, the researchers found that ChatGPT scored higher on this
test than humans did. (from
Wei) Of course, so Wei, “this does not necessarily translate into ChatGPT being
emotionally intelligent or empathetic” (a capability that wasn’t tested), nor
does it show that it has a “conversational capability in sensing and
interacting with the emotions of others.”
Nonetheless, steps into the direction of a
conversational capability have already been taken, as another research shows. One of the
problems when being in contact with other persons on the internet often is that
we don’t see them. It’s a problem because seeing others makes it possible to
read their emotions from their faces. Just the absence of face-to-face contacts
makes that some people are ruder when dealing with internet partners than when they
would have been in real-life contact with those persons. As such, contact via a
screen is not the same as a real personal contact. Now the study just mentioned
developed “an AI-in-the-loop agent” (called Hailey) “that provides just-in-time
feedback to help participants who provide support (peer supporters) respond
more empathically to those seeking help (support seekers).” Using Hailey led to
a substantial increase in feeling empathy by the peer supporters and expressing
this empathy in their contacts with support seekers. So, Hailey did not only help
peer supporters to recognize emotions in support seekers but also helped them responding
in the right way by advising how to respond. “Overall,” so
Dr. Wei, “this study represents promising and innovative research that
demonstrates how a human-AI collaboration can allow people to feel more
confident about providing support.” But for this we need AI systems that can
recognize emotions and then respond in the right way.
All this can be seen as first steps toward a world with conscious AI systems
that can be characterized as virtual humans that apparently behave like real
humans. Another
study, discussed
by Marilynn Wei, shows that such AI systems are no longer fiction but on
the way to become fact. In such a world, empathy and social connection are within
reach of AI systems. Once AI systems behave like humans, humans tend to see
them as humans (see the article by Marilynn Wei just mentioned). It’s a bit
like the famous theorem
by W.I. Thomas: If men define situations as real, they are real in their
consequences. Virtuality and reality intermingle, and the difference between men
and machines tends to disappear. Nevertheless, behaving like humans is not the
same as being human, for Chalmer’s
hard problem still stands: Even if
an AI system shows behaviour that is characteristic of having consciousness, we
still don’t know whether it really has consciousness. It is still possible that
the AI system is a zombie in the philosophical sense, because it shows
consciousness-related behaviour but doesn’t have the consciousness-related subjective
experience. Who cares?
No comments:
Post a Comment