Algorithms,
Journal Year:
2025,
Volume and Issue:
18(2), P. 86 - 86
Published: Feb. 5, 2025
Objective
(1):
To
develop
and
validate
a
machine
learning
(ML)
model
using
radiomic
features
(RFs)
extracted
from
[18F]FDG
PET-CT
to
predict
abdominal
aortic
aneurysm
(AAA)
growth
rate.
Methods
(2):
This
retrospective
study
included
98
internal
55
external
AAA
patients
undergoing
PET-CT.
RFs
were
manual
segmentations
of
AAAs
PyRadiomics.
Recursive
feature
elimination
(RFE)
reduced
for
optimisation.
A
multi-layer
perceptron
(MLP)
was
developed
prediction
compared
against
Random
Forest
(RF),
XGBoost,
Support
Vector
Machine
(SVM).
Accuracy
evaluated
via
cross-validation,
with
uncertainty
quantified
dropout
(MLP),
standard
deviation
95%
intervals
(XGBoost).
External
validation
used
independent
data
two
centres.
Ground
truth
rates
calculated
serial
ultrasound
(US)
measurements
or
CT
volumes.
Results
(3):
From
93
initial
RFs,
29
remained
after
RFE.
The
MLP
achieved
an
MAE
±
SEM
1.35
3.2e−4
mm/year
the
full
set
2.5e−4
yielded
1.8
8.9e−8
mm/year.
RF,
SVM
models
produced
comparable
accuracies
internally
(1.4–1.5
mm/year)
but
showed
higher
errors
during
(1.9–1.97
mm/year).
demonstrated
across
all
datasets.
Conclusions
(4):
An
leveraging
radiomics
accurately
predicted
generalised
well
data.
In
future,
more
sophisticated
stratification
could
guide
individualised
patient
care,
facilitating
risk-tailored
management
AAAs.
BMJ,
Journal Year:
2024,
Volume and Issue:
unknown, P. e078378 - e078378
Published: April 16, 2024
The
TRIPOD
(Transparent
Reporting
of
a
multivariable
prediction
model
for
Individual
Prognosis
Or
Diagnosis)
statement
was
published
in
2015
to
provide
the
minimum
reporting
recommendations
studies
developing
or
evaluating
performance
model.
Methodological
advances
field
have
since
included
widespread
use
artificial
intelligence
(AI)
powered
by
machine
learning
methods
develop
models.
An
update
is
thus
needed.
TRIPOD+AI
provides
harmonised
guidance
studies,
irrespective
whether
regression
modelling
been
used.
new
checklist
supersedes
checklist,
which
should
no
longer
be
This
article
describes
development
and
presents
expanded
27
item
with
more
detailed
explanation
each
recommendation,
Abstracts
checklist.
aims
promote
complete,
accurate,
transparent
that
evaluate
its
performance.
Complete
will
facilitate
study
appraisal,
evaluation,
implementation.
BMJ,
Journal Year:
2024,
Volume and Issue:
unknown, P. e074820 - e074820
Published: Jan. 15, 2024
External
validation
studies
are
an
important
but
often
neglected
part
of
prediction
model
research.
In
this
article,
the
second
in
a
series
on
evaluation,
Riley
and
colleagues
explain
what
external
study
entails
describe
key
steps
involved,
from
establishing
high
quality
dataset
to
evaluating
model's
predictive
performance
clinical
usefulness.
BMJ,
Journal Year:
2024,
Volume and Issue:
unknown, P. e074821 - e074821
Published: Jan. 22, 2024
An
external
validation
study
evaluates
the
performance
of
a
prediction
model
in
new
data,
but
many
these
studies
are
too
small
to
provide
reliable
answers.
In
third
article
their
series
on
evaluation,
Riley
and
colleagues
describe
how
calculate
sample
size
required
for
studies,
propose
avoid
rules
thumb
by
tailoring
calculations
setting
at
hand.
BMJ,
Journal Year:
2024,
Volume and Issue:
unknown, P. e078276 - e078276
Published: Sept. 3, 2024
Predicting
future
outcomes
of
patients
is
essential
to
clinical
practice,
with
many
prediction
models
published
each
year.
Empirical
evidence
suggests
that
studies
often
have
severe
methodological
limitations,
which
undermine
their
usefulness.
This
article
presents
a
step-by-step
guide
help
researchers
develop
and
evaluate
model.
The
covers
best
practices
in
defining
the
aim
users,
selecting
data
sources,
addressing
missing
data,
exploring
alternative
modelling
options,
assessing
model
performance.
steps
are
illustrated
using
an
example
from
relapsing-remitting
multiple
sclerosis.
Comprehensive
R
code
also
provided.
BMC Medical Informatics and Decision Making,
Journal Year:
2025,
Volume and Issue:
25(1)
Published: Jan. 6, 2025
Abstract
Background
Machine
learning
(ML)
is
increasingly
used
to
predict
clinical
deterioration
in
intensive
care
unit
(ICU)
patients
through
scoring
systems.
Although
promising,
such
algorithms
often
overfit
their
training
cohort
and
perform
worse
at
new
hospitals.
Thus,
external
validation
a
critical
–
but
frequently
overlooked
step
establish
the
reliability
of
predicted
risk
scores
translate
them
into
practice.
We
systematically
reviewed
how
regularly
ML-based
performed
performance
changed
data.
Methods
searched
MEDLINE,
Web
Science,
arXiv
for
studies
using
ML
ICU
from
routine
included
primary
research
published
English
before
December
2023.
summarised
many
were
externally
validated,
assessing
differences
over
time,
by
outcome,
data
source.
For
validated
studies,
we
evaluated
change
area
under
receiver
operating
characteristic
(AUROC)
attributable
linear
mixed-effects
models.
Results
572
which
84
(14.7%)
increasing
23.9%
Validated
made
disproportionate
use
open-source
data,
with
two
well-known
US
datasets
(MIMIC
eICU)
accounting
83.3%
studies.
On
average,
AUROC
was
reduced
-0.037
(95%
CI
-0.052
-0.027)
more
than
0.05
reduction
49.5%
Discussion
External
validation,
although
increasing,
remains
uncommon.
Performance
generally
lower
questioning
some
recently
proposed
scores.
Interpretation
results
challenged
an
overreliance
on
same
few
datasets,
implicit
case
mix,
exclusive
AUROC.
Nature Communications,
Journal Year:
2025,
Volume and Issue:
16(1)
Published: Jan. 8, 2025
Abstract
Infertility
affects
one-in-six
couples,
often
necessitating
in
vitro
fertilization
treatment
(IVF).
IVF
generates
complex
data,
which
can
challenge
the
utilization
of
full
richness
data
during
decision-making,
leading
to
reliance
on
simple
‘rules-of-thumb’.
Machine
learning
techniques
are
well-suited
analyzing
provide
data-driven
recommendations
improve
decision-making.
In
this
multi-center
study
(
n
=
19,082
treatment-naive
female
patients),
including
11
European
centers,
we
harnessed
explainable
artificial
intelligence
identify
follicle
sizes
that
contribute
most
relevant
downstream
clinical
outcomes.
We
found
intermediately-sized
follicles
were
important
number
mature
oocytes
subsequently
retrieved.
Maximizing
proportion
by
end
ovarian
stimulation
was
associated
with
improved
live
birth
rates.
Our
suggests
larger
mean
sizes,
especially
those
>18
mm,
premature
progesterone
elevation
and
a
negative
impact
rates
fresh
embryo
transfer.
These
highlight
potential
computer
technologies
aid
personalization
optimize
outcomes
pending
future
prospective
validation.
BMJ,
Journal Year:
2025,
Volume and Issue:
unknown, P. e082505 - e082505
Published: March 24, 2025
The
Prediction
model
Risk
Of
Bias
ASsessment
Tool
(PROBAST)
is
used
to
assess
the
quality,
risk
of
bias,
and
applicability
prediction
models
or
algorithms
model/algorithm
studies.
Since
PROBAST's
introduction
in
2019,
much
progress
has
been
made
methodology
for
modelling
use
artificial
intelligence,
including
machine
learning,
techniques.
An
update
PROBAST-2019
thus
needed.
This
article
describes
development
PROBAST+AI.
PROBAST+AI
consists
two
distinctive
parts:
evaluation.
For
development,
users
quality
using
16
targeted
signalling
questions.
evaluation,
bias
18
Both
parts
contain
four
domains:
participants
data
sources,
predictors,
outcome,
analysis.
Applicability
rated
outcome
domains.
may
replace
original
PROBAST
tool
allows
all
key
stakeholders
(eg,
developers,
AI
companies,
researchers,
editors,
reviewers,
healthcare
professionals,
guideline
policy
organisations)
examine
any
type
sector,
irrespective
whether
regression
techniques
are
used.
Computers in Biology and Medicine,
Journal Year:
2025,
Volume and Issue:
191, P. 110183 - 110183
Published: April 14, 2025
Pre-eclampsia
(PE)
contributes
to
more
than
one-fourth
of
all
maternal
deaths
and
half
a
million
newborn
worldwide
every
year.
Early
screening
interventions
can
reduce
PE
incidence
related
complications.
We
aim
1)
temporally
validate
three
existing
models
(two
machine
learning
(ML)
one
logistic
regression)
developed
in
the
same
region
2)
compare
performances
validated
ML
with
regression
model
prediction.
This
work
addresses
gap
literature
by
undertaking
comprehensive
evaluation
risk
prediction
models,
which
is
an
important
step
advancing
this
field.
obtained
dataset
routinely
collected
antenatal
data
from
maternity
hospitals
South-East
Melbourne,
Australia,
extracted
between
July
2021
December
2022.
models:
extreme
gradient
boosting
(XGBoost,
'model
1'),
random
forest
('model
2')
3').
Area
under
receiver-operating
characteristic
(ROC)
curve
(AUC)
was
evaluated
discrimination
performance,
calibration
assessed.
The
AUCs
were
compared
using
'bootstrapping'
test.
temporal
consisted
12,549
singleton
pregnancies,
431
(3.43
%,
95
%
confidence
interval
(CI)
3.13-3.77)
PE.
characteristics
similar
original
development
dataset.
XGBoost
1'
3'
exhibited
performance
AUC
0.75
(95
CI
0.73-0.78)
0.76
0.74-0.78),
respectively.
2'
showed
0.71
0.69-0.74).
Model
3
perfect
slope
1.02
0.92-1.12).
Models
1
2
1.15
1.03-1.28)
0.62
0.54-0.70),
Compared
stable
whereas
significantly
lower
performance.
better
clinical
net
benefits
22
threshold
probabilities
default
strategies.
During
validation
performance;
however,
both
did
not
outperform
model.
To
facilitate
insights
into
interpretability
deployment,
could
be
integrated
routine
practice
as
first-step
two-stage
approach
identify
higher-risk
woman
for
further
second
stage
accurate
The Lancet Global Health,
Journal Year:
2024,
Volume and Issue:
12(8), P. e1343 - e1358
Published: July 17, 2024
Cardiovascular
diseases
remain
the
number
one
cause
of
death
globally.
disease
risk
scores
are
an
integral
tool
in
primary
prevention,
being
used
to
identify
individuals
at
highest
and
guide
assignment
preventive
interventions.
Available
differ
substantially
terms
population
sample
data
sources
for
their
derivation
and,
consequently,
absolute
risks
they
assign
individuals.
Differences
cardiovascular
epidemiology
between
populations
contributing
development
scores,
target
which
applied,
can
result
overestimation
or
underestimation
individuals,
poorly
informed
clinical
decisions.
Given
wide
plethora
available,
identification
appropriate
score
a
be
challenging.
This
Review
provides
up-to-date
overview
guideline-recommended
from
global,
regional,
national
contexts,
evaluates
comparative
characteristics
qualities,
guidance
on
selection
score.