A Comparative Study on the Question-Answering Proficiency of Artificial Intelligence Models in Bladder-Related Conditions: An Evaluation of Gemini and ChatGPT 4.o
Medical Records,
Journal Year:
2025,
Volume and Issue:
7(1), P. 201 - 205
Published: Jan. 10, 2025
Aim:
The
rapid
evolution
of
artificial
intelligence
(AI)
has
revolutionized
medicine,
with
tools
like
ChatGPT
and
Google
Gemini
enhancing
clinical
decision-making.
ChatGPT's
advancements,
particularly
GPT-4,
show
promise
in
diagnostics
education.
However,
variability
accuracy
limitations
complex
scenarios
emphasize
the
need
for
further
evaluation
these
models
medical
applications.
This
study
aimed
to
assess
agreement
between
4.o
AI
identifying
bladder-related
conditions,
including
neurogenic
bladder,
vesicoureteral
reflux
(VUR),
posterior
urethral
valve
(PUV).
Material
Method:
study,
conducted
October
2024,
compared
AI's
on
51
questions
about
VUR,
PUV.
Questions,
randomly
selected
from
pediatric
surgery
urology
materials,
were
evaluated
using
metrics
statistical
analysis,
highlighting
models'
performance
agreement.
Results:
demonstrated
similar
across
PUV
questions,
true
response
rates
66.7%
68.6%,
respectively,
no
statistically
significant
differences
(p>0.05).
Combined
all
topics
was
67.6%.
Strong
inter-rater
reliability
(κ=0.87)
highlights
their
Conclusion:
comparable
ChatGPT-4.o
key
performance.
Language: Английский
AI‐driven simplification of surgical reports in gynecologic oncology: A potential tool for patient education
M Riedel,
No information about this author
Bastian Meyer,
No information about this author
R.D. Rubens
No information about this author
et al.
Acta Obstetricia Et Gynecologica Scandinavica,
Journal Year:
2025,
Volume and Issue:
unknown
Published: May 14, 2025
Abstract
Introduction
The
emergence
of
large
language
models
heralds
a
new
chapter
in
natural
processing,
with
immense
potential
for
improving
medical
care
and
especially
oncology.
One
recent
publicly
available
example
is
Generative
Pretraining
Transformer
4
(GPT‐4).
Our
objective
was
to
evaluate
its
ability
rephrase
original
surgical
reports
into
simplified
versions
that
are
more
comprehensible
patients.
Specifically,
we
aimed
investigate
discuss
the
potential,
limitations,
associated
risks
using
these
patient
education
information
gynecologic
Material
Methods
We
tasked
GPT‐4
generating
from
n
=
20
reports.
Patients
were
provided
both
their
report
corresponding
version
generated
by
GPT‐4.
Alongside
reports,
patients
received
questionnaires
designed
facilitate
comparative
assessment
between
Furthermore,
clinical
experts
evaluated
artificial
intelligence
(AI)‐generated
regard
accuracy
quality.
Results
significantly
improved
our
patients'
understanding,
particularly
procedure,
outcome,
risks.
However,
despite
being
accessible
relevant,
highlighted
concerns
about
lack
precision.
Conclusions
Advanced
like
can
transform
unedited
improve
clarity
procedure
outcomes.
It
offers
considerable
promise
enhancing
education.
precision
underscore
need
rigorous
oversight
safely
integrate
AI
Over
medium
term,
AI‐generated,
reports—and
other
records—could
be
effortlessly
integrated
standard
automated
postoperative
digital
discharge
systems.
Language: Английский