Cognitive vs. emotional empathy: exploring their impact on user outcomes in health-assistant chatbots
Behaviour and Information Technology,
Journal Year:
2025,
Volume and Issue:
unknown, P. 1 - 16
Published: March 6, 2025
Language: Английский
Artificial intelligence and psychoanalysis: is it time for psychoanalyst.AI?
Frontiers in Psychiatry,
Journal Year:
2025,
Volume and Issue:
16
Published: April 7, 2025
The
current
development
of
artificial
intelligences
(AI)
is
leading
to
major
transformations
within
society.
In
this
context,
we
observe
how
some
these
AIs
are
spontaneously
used
by
individuals
as
confidants,
and
even
romantic
partners.
emergence
such
relationships
with
raises
questions
about
their
integration
in
psychiatry
the
possibility
developing
"digital
therapists".
regard,
highlight
four
key
elements
(accessibility
availability;
confidentiality;
knowledge;
memory)
compare
what
an
AI
offers
comparison
a
human
therapist.
We
also
discuss
results
studies
that
have
already
investigated
use
psychotherapy,
particularly
fields
depression
anxiety.
then
propose
reflect
more
specifically
on
creating
"psychoanalyst.AI,"
which
leads
us
examine
therapeutic
relationship
(transference,
free
association,
play,
dreams,
reflexivity,
narrativity)
AI.
conclusion,
offer
reflections
relevance
considering
"therapeutic
artifact,"
while
taking
into
account
ethical
issues
raised
settings.
Language: Английский
ChatGPT's role in sleep health: Informative or misleading
Sleep Health,
Journal Year:
2025,
Volume and Issue:
unknown
Published: April 1, 2025
Language: Английский
Leveraging AI-Generated Emotional Self-Voice to Nudge People towards their Ideal Selves
Published: April 24, 2025
Language: Английский
Impact of Large Language Model–Based AI Tools on Physician–Patient Communication: A Systematic Review and Meta-Analysis (Preprint)
Sven Richter,
No information about this author
Clara Buszello,
No information about this author
Markus Prem
No information about this author
et al.
Published: May 11, 2025
BACKGROUND
Recent
advances
in
large
language
models
(LLMs)
such
as
GPT-3/4
have
spurred
development
of
AI
chatbots
and
advisory
tools
medicine.
These
systems
are
posited
to
assist
or
augment
physician–patient
communication,
potentially
improving
empathy,
clarity,
responsiveness.
However,
their
actual
impact
on
communication
outcomes
remains
uncertain.
OBJECTIVE
To
systematically
review
meta-analyze
peer-reviewed
studies
(2020–2025)
evaluating
how
LLM-based
interventions
affect
including
trust,
patient
understanding.
METHODS
Following
PRISMA
2020
guidelines,
we
searched
PubMed/MEDLINE
for
published
from
2025
examining
LLM
chatbot
applications
clinical
contexts.
Eligible
designs
included
randomized,
observational,
cross-sectional,
qualitative
studies.
Two
reviewers
independently
screened
titles/abstracts,
assessed
full
texts,
extracted
data
study
design,
population,
type,
measures,
outcomes.
We
conducted
a
synthesis
random-effects
meta-analysis,
reporting
pooled
standardized
mean
differences
(SMD)
odds
ratios
(OR)
with
95%
confidence
intervals
(CI).
RESULTS
From
312
records,
10
(N=10)
were
included,
all
quantitative
predominantly
cross-sectional.
Populations
ranged
patients
chronic
conditions
healthcare
professionals
laypersons.
Outcomes
empathy
(7
studies),
clarity/information
quality
(6),
satisfaction
usefulness
(4),
trust
perceptions
(2).
In
six
direct
comparisons
AI-
versus
physician-generated
responses,
LLMs
rated
significantly
higher
five
One
found
replies
judged
empathetic
45.1%
cases
4.6%
physician
(OR
~9.8,
P<.001).
Similarly,
ChatGPT-4
answers
scored
5-point
scale
than
human-written
responses
(mean
4.18
vs
2.70,
neurology
showed
scores
(CARE
+1.38,
P<.01)
ChatGPT
answers.
Only
one
no
significant
difference.
content
was
also
longer
more
information-rich,
patient-perceived
clarity
On
the
other
hand,
GPT-4
simplified
pathology
reports,
increasing
comprehension
(7.98
5.23/10,
P<.001)
reducing
consultation
time
by
70%.
sometimes
less
concise
readable
low-literacy
patients.
analyses
(4
studies,
n=2,604),
positive
effect
(SMD
+1.05,
CI
0.45–1.65)
improved
understanding
+0.82,
0.30–1.34).
Patient
results
mixed.
No
directly
long-term
trust.
CONCLUSIONS
Current
evidence
suggests
can
enhance
producing
empathetic,
detailed,
understandable
responses.
improvements
may
positively
influence
experience
engagement.
generate
overly
lengthy
occasionally
inaccurate
advice,
emphasizing
need
oversight.
While
meta-analytic
findings
promising,
robust
controlled
trials
needed
confirm
benefits,
assess
outcomes,
define
optimal
integration
strategies.
Language: Английский