AI-Driven Innovations in Alzheimer's Disease: Integrating Early Diagnosis, Personalized Treatment, and Prognostic Modelling
Ageing Research Reviews,
Journal Year:
2024,
Volume and Issue:
unknown, P. 102497 - 102497
Published: Sept. 1, 2024
Language: Английский
Improving large language model applications in biomedicine with retrieval-augmented generation: a systematic review, meta-analysis, and clinical development guidelines
Journal of the American Medical Informatics Association,
Journal Year:
2025,
Volume and Issue:
unknown
Published: Jan. 15, 2025
The
objectives
of
this
study
are
to
synthesize
findings
from
recent
research
retrieval-augmented
generation
(RAG)
and
large
language
models
(LLMs)
in
biomedicine
provide
clinical
development
guidelines
improve
effectiveness.
We
conducted
a
systematic
literature
review
meta-analysis.
report
was
created
adherence
the
Preferred
Reporting
Items
for
Systematic
Reviews
Meta-Analyses
2020
analysis.
Searches
were
performed
3
databases
(PubMed,
Embase,
PsycINFO)
using
terms
related
"retrieval
augmented
generation"
"large
model,"
articles
published
2023
2024.
selected
studies
that
compared
baseline
LLM
performance
with
RAG
performance.
developed
random-effect
meta-analysis
model,
odds
ratio
as
effect
size.
Among
335
studies,
20
included
review.
pooled
size
1.35,
95%
confidence
interval
1.19-1.53,
indicating
statistically
significant
(P
=
.001).
reported
tasks,
LLMs,
retrieval
sources
strategies,
well
evaluation
methods.
Building
on
our
review,
we
Guidelines
Unified
Implementation
Development
Enhanced
Applications
Clinical
Settings
inform
applications
RAG.
Overall,
implementation
showed
1.35
increase
LLMs.
Future
should
focus
(1)
system-level
enhancement:
combination
agent,
(2)
knowledge-level
deep
integration
knowledge
into
LLM,
(3)
integration-level
integrating
systems
within
electronic
health
records.
Language: Английский
Generative Large Language Models in Electronic Health Records for Patient Care Since 2023: A Systematic Review
medRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2024,
Volume and Issue:
unknown
Published: Aug. 12, 2024
Background:
Generative
Large
language
models
(LLMs)
represent
a
significant
advancement
in
natural
processing,
achieving
state-of-the-art
performance
across
various
tasks.
However,
their
application
clinical
settings
using
real
electronic
health
records
(EHRs)
is
still
rare
and
presents
numerous
challenges.
Objective:
This
study
aims
to
systematically
review
the
use
of
generative
LLMs,
effectiveness
relevant
techniques
patient
care-related
topics
involving
EHRs,
summarize
challenges
faced,
suggest
future
directions.
Methods:
A
Boolean
search
for
peer-reviewed
articles
was
conducted
on
May
19th,
2024
PubMed
Web
Science
include
research
published
since
2023,
which
one
month
after
release
ChatGPT.
The
results
were
deduplicated.
Multiple
reviewers,
including
biomedical
informaticians,
computer
scientists,
physician,
screened
publications
eligibility
data
extraction.
Only
studies
utilizing
LLMs
analyze
EHR
included.
We
summarized
prompt
engineering,
fine-tuning,
multimodal
data,
evaluation
matrices.
Additionally,
we
identified
current
applying
as
reported
by
included
proposed
Results:
initial
6,328
unique
studies,
with
76
screening.
Of
these,
67
(88.2%)
employed
zero-shot
prompting,
five
them
100%
accuracy
specific
Nine
used
advanced
prompting
strategies;
four
tested
these
strategies
experimentally,
finding
that
engineering
improved
performance,
noting
non-linear
relationship
between
number
examples
improvement.
Eight
explored
fine-tuning
all
improvements
tasks,
but
three
noted
potential
degradation
certain
two
utilized
LLM-based
decision-making
enabled
accurate
disease
diagnosis
prognosis.
55
different
metrics
22
purposes,
such
correctness,
completeness,
conciseness.
Two
investigated
LLM
bias,
detecting
no
bias
other
male
patients
received
more
appropriate
suggestions.
Six
hallucinations,
fabricating
names
structured
thyroid
ultrasound
reports.
Additional
not
limited
impersonal
tone
consultations,
made
uncomfortable,
difficulty
had
understanding
responses.
Conclusion:
Our
indicates
few
have
computational
enhance
performance.
diverse
highlight
need
standardization.
currently
cannot
replace
physicians
due
Language: Английский
Assessing the Potential of USMLE-Like Exam Questions Generated by GPT-4
medRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2023,
Volume and Issue:
unknown
Published: April 28, 2023
The
United
States
Medical
Licensing
Examination
(USMLE)
is
a
critical
step
in
assessing
the
competence
of
future
physicians,
yet
process
creating
exam
questions
and
study
materials
both
time-consuming
costly.
While
Large
Language
Models
(LLMs),
such
as
OpenAI’s
GPT-4,
have
demonstrated
proficiency
answering
medical
questions,
their
potential
generating
remains
underexplored.
This
presents
QUEST-AI,
novel
system
that
utilizes
LLMs
to
(1)
generate
USMLE-style
(2)
identify
flag
incorrect
(3)
correct
errors
flagged
questions.
We
evaluated
this
system’s
output
by
constructing
test
set
50
LLM-generated
mixed
with
human-generated
conducting
two-part
assessment
three
physicians
two
students.
assessors
attempted
distinguish
between
LLM
validity
content.
A
majority
generated
QUEST-AI
were
deemed
valid
panel
clinicians,
strong
correlations
performance
on
pioneering
application
education
could
significantly
increase
ease
efficiency
developing
content,
offering
cost-effective
accessible
alternative
for
preparation.
Language: Английский
Identification of an ANCA-associated vasculitis cohort using deep learning and electronic health records
Liqin Wang,
No information about this author
John Laurentiev,
No information about this author
Claire Cook
No information about this author
et al.
International Journal of Medical Informatics,
Journal Year:
2025,
Volume and Issue:
196, P. 105797 - 105797
Published: Jan. 18, 2025
Language: Английский
AI-powered model for accurate prediction of MCI-to-AD progression
Acta Pharmaceutica Sinica B,
Journal Year:
2025,
Volume and Issue:
unknown
Published: Feb. 1, 2025
Language: Английский
Evaluating local open-source large language models for data extraction from unstructured reports on mechanical thrombectomy in patients with ischemic stroke
Aymen Meddeb,
No information about this author
Philipe Ebert,
No information about this author
Keno K. Bressem
No information about this author
et al.
Journal of NeuroInterventional Surgery,
Journal Year:
2024,
Volume and Issue:
unknown, P. jnis - 022078
Published: Aug. 2, 2024
Background
A
study
was
undertaken
to
assess
the
effectiveness
of
open-source
large
language
models
(LLMs)
in
extracting
clinical
data
from
unstructured
mechanical
thrombectomy
reports
patients
with
ischemic
stroke
caused
by
a
vessel
occlusion.
Methods
We
deployed
local
LLMs
extract
points
free-text
procedural
who
underwent
between
September
2020
and
June
2023
our
institution.
The
external
dataset
obtained
second
university
hospital
comprised
consecutive
cases
treated
March
2024.
Ground
truth
labeling
facilitated
human-in-the-loop
(HITL)
approach,
time
metrics
recorded
for
both
automated
manual
extractions.
tested
three
models—Mixtral,
Qwen,
BioMistral—assessing
their
performance
on
precision,
recall,
F1
score
across
15
categories
such
as
National
Institute
Health
Stroke
Scale
(NIHSS)
scores,
occluded
vessels,
medication
details.
Results
included
1000
primary
institution
50
secondary
Mixtral
showed
highest
achieving
0.99
first
series
extraction
0.69
identification
within
internal
dataset.
In
dataset,
precision
ranged
1.00
NIHSS
scores
0.70
vessels.
Qwen
moderate
high
0.85
low
0.28
BioMistral
had
broadest
range
0.81
times
0.14
HITL
approach
yielded
an
average
savings
65.6%
per
case,
variations
45.95%
79.56%.
Conclusion
This
highlights
potential
using
medical
reports.
Incorporating
annotations
enhances
also
ensures
reliability
extracted
data.
methodology
presents
scalable
privacy-preserving
option
that
can
significantly
support
documentation
research
endeavors.
Language: Английский
SCD-Tron: Leveraging Large Clinical Language Model for Early Detection of Cognitive Decline from Electronic Health Records
Hao Guan,
No information about this author
John Novoa-Laurentiev,
No information about this author
Zhou Li
No information about this author
et al.
medRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2024,
Volume and Issue:
unknown
Published: Nov. 2, 2024
Background:
Early
detection
of
cognitive
decline
during
the
preclinical
stage
Alzheimer's
disease
and
related
dementias
(AD/ADRD)
is
crucial
for
timely
intervention
treatment.
Clinical
notes
in
electronic
health
record
contain
valuable
information
that
can
aid
early
identification
decline.
In
this
study,
we
utilize
advanced
large
clinical
language
models,
fine-tuned
on
notes,
to
improve
Methods:
We
collected
from
2,166
patients
spanning
4
years
preceding
their
initial
mild
impairment
(MCI)
diagnosis
Enterprise
Data
Warehouse
Mass
General
Brigham.
To
train
model,
developed
CD-Tron,
built
upon
a
model
was
finetuned
using
4,949
expert-labeled
note
sections.
For
evaluation,
trained
applied
1,996
independent
sections
assess
its
performance
real-world
unstructured
data.
Additionally,
used
explainable
AI
techniques,
specifically
SHAP
values
(SHapley
Additive
exPlanations),
interpret
model's
predictions
provide
insight
into
most
influential
features.
Error
analysis
also
facilitated
further
analyze
prediction.
Results:
CD-Tron
significantly
outperforms
baseline
achieving
notable
improvements
precision,
recall,
AUC
metrics
detecting
(CD).
Tested
many
demonstrated
high
sensitivity
with
only
one
false
negative,
applications
prioritizing
accurate
CD
detection.
SHAP-based
interpretability
highlighted
key
textual
features
contributing
predictions,
supporting
transparency
clinician
understanding.
Conclusion:
offers
novel
approach
by
applying
models
free-text
EHR
Pretrained
it
accurately
identifies
integrates
interpretability,
enhancing
predictions.
Language: Английский