Current Opinion in Ophthalmology,
Год журнала:
2024,
Номер
unknown
Опубликована: Сен. 26, 2024
Purpose
of
review
Last
year
marked
the
development
first
foundation
model
in
ophthalmology,
RETFound,
setting
stage
for
generalizable
medical
artificial
intelligence
(GMAI)
that
can
adapt
to
novel
tasks.
Additionally,
rapid
advancements
large
language
(LLM)
technology,
including
models
such
as
GPT-4
and
Gemini,
have
been
tailored
specialization
evaluated
on
clinical
scenarios
with
promising
results.
This
explores
opportunities
challenges
further
these
technologies.
Recent
findings
RETFound
outperforms
traditional
deep
learning
specific
tasks,
even
when
only
fine-tuned
small
datasets.
LMMs
like
Med-Gemini
Medprompt
perform
better
than
out-of-the-box
ophthalmology
However,
there
is
still
a
significant
deficiency
ophthalmology-specific
multimodal
models.
gap
primarily
due
substantial
computational
resources
required
train
limitations
high-quality
Summary
Overall,
present
but
face
challenges,
particularly
need
high-quality,
standardized
datasets
training
specialization.
Although
has
focused
vision
models,
greatest
lie
advancing
which
more
closely
mimic
capabilities
clinicians.
Clinical Chemistry and Laboratory Medicine (CCLM),
Год журнала:
2025,
Номер
unknown
Опубликована: Апрель 18, 2025
Abstract
Objectives
Accurate
medical
laboratory
reports
are
essential
for
delivering
high-quality
healthcare.
Recently,
advanced
artificial
intelligence
models,
such
as
those
in
the
ChatGPT
series,
have
shown
considerable
promise
this
domain.
This
study
assessed
performance
of
specific
GPT
models-namely,
4o,
o1,
and
o1
mini-in
identifying
errors
within
providing
treatment
recommendations.
Methods
In
retrospective
study,
86
Nucleic
acid
test
report
seven
upper
respiratory
tract
pathogens
were
compiled.
There
285
from
four
common
error
categories
intentionally
randomly
introduced
into
generated
incorrected
reports.
models
tasked
with
detecting
these
errors,
using
three
senior
scientists
(SMLS)
interns
(MLI)
control
groups.
Additionally,
generating
accurate
reliable
recommendations
following
positive
outcomes
based
on
corrected
χ2
tests,
Kruskal-Wallis
Wilcoxon
tests
used
statistical
analysis
where
appropriate.
Results
comparison
SMLS
or
MLI,
accurately
detected
types,
average
detection
rates
88.9
%(omission),
91.6
%
(time
sequence),
91.7
(the
same
individual
acted
both
inspector
reviewer).
However,
rate
result
input
format
by
was
only
51.9
%,
indicating
a
relatively
poor
aspect.
exhibited
substantial
to
almost
perfect
agreement
total
(kappa
[min,
max]:
0.778,
0.837).
between
MLI
moderately
lower
0.632,
0.696).
When
it
comes
reading
all
reports,
showed
obviously
reduced
time
compared
(all
p<0.001).
Notably,
our
also
found
GPT-o1
mini
model
had
better
consistency
identification
than
model,
which
that
GPT-4o
model.
The
pairwise
comparisons
model’s
outputs
across
repeated
runs
0.912,
0.996).
GPT-o1(all
significantly
outperformed
p<0.0001).
Conclusions
capability
some
accuracy
reliability
competent,
especially,
potentially
reducing
work
hours
enhancing
clinical
decision-making.
Current Opinion in Ophthalmology,
Год журнала:
2024,
Номер
35(5), С. 391 - 402
Опубликована: Май 30, 2024
Large
language
models
(LLMs)
are
rapidly
entering
the
landscape
of
medicine
in
areas
from
patient
interaction
to
clinical
decision-making.
This
review
discusses
evolving
role
LLMs
ophthalmology,
focusing
on
their
current
applications
and
future
potential
enhancing
ophthalmic
care.
Frontiers in Digital Health,
Год журнала:
2024,
Номер
6
Опубликована: Ноя. 5, 2024
Artificial
Intelligence
(AI)
has
the
potential
to
revolutionize
medical
training,
diagnostics,
treatment
planning,
and
healthcare
delivery
while
also
bringing
challenges
such
as
data
privacy,
risk
of
technological
overreliance,
preservation
critical
thinking.
This
manuscript
explores
impact
AI
Machine
Learning
(ML)
on
interactions,
focusing
faculty,
students,
clinicians,
patients.
ML's
early
inclusion
in
curriculum
will
support
student-centered
learning;
however,
all
stakeholders
require
specialized
training
bridge
gap
between
practice
innovation.
underscores
importance
education
ethical
responsible
use
emphasizing
collaboration
maximize
its
benefits.
calls
for
a
re-evaluation
interpersonal
relationships
within
improve
overall
quality
care
safeguard
welfare
by
leveraging
AI's
strengths
managing
risks.
Ophthalmology and Therapy,
Год журнала:
2024,
Номер
unknown
Опубликована: Ноя. 8, 2024
Cataracts
are
a
significant
cause
of
blindness.
While
individuals
frequently
turn
to
the
Internet
for
medical
advice,
distinguishing
reliable
information
can
be
challenging.
Large
language
models
(LLMs)
have
attracted
attention
generating
accurate,
human-like
responses
that
may
used
consultation.
However,
comprehensive
assessment
LLMs'
accuracy
within
specific
domains
is
still
lacking.
BMC Medical Informatics and Decision Making,
Год журнала:
2024,
Номер
24(1)
Опубликована: Ноя. 26, 2024
The
large
language
models
(LLMs),
most
notably
ChatGPT,
released
since
November
30,
2022,
have
prompted
shifting
attention
to
their
use
in
medicine,
particularly
for
supporting
clinical
decision-making.
However,
there
is
little
consensus
the
medical
community
on
how
LLM
performance
contexts
should
be
evaluated.
We
performed
a
literature
review
of
PubMed
identify
publications
between
December
1,
and
April
2024,
that
discussed
assessments
LLM-generated
diagnoses
or
treatment
plans.
selected
108
relevant
articles
from
analysis.
frequently
used
LLMs
were
GPT-3.5,
GPT-4,
Bard,
LLaMa/Alpaca-based
models,
Bing
Chat.
five
criteria
scoring
outputs
"accuracy",
"completeness",
"appropriateness",
"insight",
"consistency".
defining
high-quality
been
consistently
by
researchers
over
past
1.5
years.
identified
high
degree
variation
studies
reported
findings
assessed
performance.
Standardized
reporting
qualitative
evaluation
metrics
assess
quality
can
developed
facilitate
research
healthcare.
Research Square (Research Square),
Год журнала:
2024,
Номер
unknown
Опубликована: Апрель 12, 2024
AbstractBackground:
The
integration
of
artificial
intelligence
(AI)
in
healthcare
education
is
inevitable.
Understanding
the
proficiency
generative
AI
different
languages
to
answer
complex
questions
crucial
for
educational
purposes.
Objective:
To
compare
performance
ChatGPT-4
and
Gemini
answering
Virology
multiple-choice
(MCQs)
English
Arabic,
while
assessing
quality
generated
content.
Methods:
Both
models’
responses
40
MCQs
were
assessed
correctness
based
on
CLEAR
tool
designed
evaluation
AI-generated
classified
into
lower
higher
cognitive
categories
revised
Bloom’s
taxonomy.
study
design
considered
METRICS
checklist
reporting
AI-based
studies
healthcare.
Results:
performed
better
compared
with
consistently
surpassing
scores.
led
80%
vs.
62.5%
65%
55%
Arabic.
For
both
models,
superior
domains
was
reported.
Conclusion:
Both
exhibited
potential
applications;
nevertheless,
their
varied
across
highlighting
importance
continued
development
ensure
effective
globally.