Triple-stage crack detection in stone masonry using YOLO-ensemble, MobileNetV2U-net, and spectral clustering
Automation in Construction,
Год журнала:
2025,
Номер
172, С. 106045 - 106045
Опубликована: Фев. 7, 2025
Язык: Английский
Bilingual Sign Language Recognition: A YOLOv11-Based Model for Bangla and English Alphabets
Journal of Imaging,
Год журнала:
2025,
Номер
11(5), С. 134 - 134
Опубликована: Апрель 27, 2025
Communication
through
sign
language
effectively
helps
both
hearing-
and
speaking-impaired
individuals
connect.
However,
there
are
problems
with
the
interlingual
communication
between
Bangla
Sign
Language
(BdSL)
English
(ASL)
due
to
absence
of
a
unified
system.
This
study
aims
introduce
detection
system
that
incorporates
these
two
languages
enhance
flow
for
those
who
use
forms
language.
developed
tested
deep
learning-based
sign-language
can
recognize
BdSL
ASL
alphabets
concurrently
in
real
time.
The
approach
uses
YOLOv11
object
architecture
has
been
trained
an
open-source
dataset
on
set
9556
images
containing
64
different
letter
signs
from
languages.
Data
preprocessing
was
applied
performance
model.
Evaluation
criteria,
including
precision,
recall,
mAP,
other
parameter
values
were
also
computed
evaluate
analysis
proposed
method
shows
precision
99.12%
average
recall
rates
99.63%
30
epochs.
studies
show
model
outperforms
current
techniques
recognition
(SLR)
be
used
communicating
assistive
technologies
human-computer
interaction
systems.
Язык: Английский
YOLOv9s-Pear: A Lightweight YOLOv9s-Based Improved Model for Young Red Pear Small-Target Recognition
Yi Shi,
Zhen Duan,
Shunhao Qing
и другие.
Agronomy,
Год журнала:
2024,
Номер
14(9), С. 2086 - 2086
Опубликована: Сен. 12, 2024
With
the
advancement
of
computer
vision
technology,
demand
for
fruit
recognition
in
agricultural
automation
is
increasing.
To
improve
accuracy
and
efficiency
recognizing
young
red
pears,
this
study
proposes
an
improved
model
based
on
lightweight
YOLOv9s,
termed
YOLOv9s-Pear.
By
constructing
a
feature-rich
diverse
image
dataset
pears
introducing
spatial-channel
decoupled
downsampling
(SCDown),
C2FUIBELAN,
YOLOv10
detection
head
(v10detect)
modules,
YOLOv9s
was
enhanced
to
achieve
efficient
small
targets
resource-constrained
environments.
Images
were
captured
at
different
times
locations
underwent
preprocessing
establish
high-quality
dataset.
For
improvements,
integrated
general
inverted
bottleneck
blocks
from
C2f
MobileNetV4
with
RepNCSPELAN4
module
form
new
C2FUIBELAN
module,
enhancing
model’s
training
speed
small-scale
object
detection.
Additionally,
SCDown
v10detect
modules
replaced
original
AConv
structures
model,
further
improving
performance.
The
experimental
results
demonstrated
that
YOLOv9s-Pear
achieved
high
while
reducing
computational
costs
parameters.
accuracy,
recall,
mean
precision,
extended
precision
0.971,
0.970,
0.991,
0.848,
respectively.
These
confirm
SCDown,
pear
tasks.
findings
not
only
provide
fast
accurate
technique
but
also
offer
reference
detecting
fruits
other
trees,
significantly
contributing
technology.
Язык: Английский
Performance Evaluation of Deep Learning Image Classification Modules in the MUN-ABSAI Ice Risk Management Architecture
Sensors,
Год журнала:
2025,
Номер
25(2), С. 326 - 326
Опубликована: Янв. 8, 2025
The
retreat
of
Arctic
sea
ice
has
opened
new
maritime
routes,
offering
faster
shipping
opportunities;
however,
these
routes
present
significant
navigational
challenges
due
to
the
harsh
conditions.
To
address
challenges,
this
paper
proposes
a
deep
learning-based
risk
management
architecture
with
multiple
modules,
including
classification,
assessment,
floe
tracking,
and
load
calculations.
A
comprehensive
dataset
15,000
images
was
created
using
public
sources
contributions
from
Canadian
Coast
Guard,
it
used
support
development
evaluation
system.
performance
YOLOv8n-cls
model
assessed
for
classification
modules
its
fast
inference
speed,
making
suitable
resource-constrained
onboard
systems.
training
were
conducted
across
platforms,
Roboflow,
Google
Colab,
Compute
Canada,
allowing
detailed
comparison
their
capabilities
in
image
preprocessing,
training,
real-time
generation.
results
demonstrate
that
Image
Classification
Module
I
achieved
validation
accuracy
99.4%,
while
II
attained
98.6%.
Inference
times
found
be
less
than
1
s
Colab
under
3
on
stand-alone
system,
confirming
architecture's
efficiency
condition
monitoring.
Язык: Английский
Real-Time American Sign Language Interpretation Using Deep Learning and Keypoint Tracking
Sensors,
Год журнала:
2025,
Номер
25(7), С. 2138 - 2138
Опубликована: Март 28, 2025
Communication
barriers
pose
significant
challenges
for
the
Deaf
and
Hard-of-Hearing
(DHH)
community,
limiting
their
access
to
essential
services,
social
interactions,
professional
opportunities.
To
bridge
this
gap,
assistive
technologies
leveraging
artificial
intelligence
(AI)
deep
learning
have
gained
prominence.
This
study
presents
a
real-time
American
Sign
Language
(ASL)
interpretation
system
that
integrates
with
keypoint
tracking
enhance
accessibility
foster
inclusivity.
By
combining
YOLOv11
model
gesture
recognition
MediaPipe
precise
hand
tracking,
achieves
high
accuracy
in
identifying
ASL
alphabet
letters
real
time.
The
proposed
approach
addresses
such
as
ambiguity,
environmental
variations,
computational
efficiency.
Additionally,
enables
users
spell
out
names
locations,
further
improving
its
practical
applications.
Experimental
results
demonstrate
attains
mean
Average
Precision
([email protected])
of
98.2%,
an
inference
speed
optimized
real-world
deployment.
research
underscores
critical
role
AI-driven
empowering
DHH
community
by
enabling
seamless
communication
interaction.
Язык: Английский
Enhanced Vehicle Identification Using YOLOv8 with Counter-Based Grouping for Improved Real-Time Performance
Algorithms for intelligent systems,
Год журнала:
2025,
Номер
unknown, С. 1 - 13
Опубликована: Янв. 1, 2025
Язык: Английский
Small Object Detection in UAV Remote Sensing Images Based on Intra-Group Multi-Scale Fusion Attention and Adaptive Weighted Feature Fusion Mechanism
Remote Sensing,
Год журнала:
2024,
Номер
16(22), С. 4265 - 4265
Опубликована: Ноя. 15, 2024
In
view
of
the
issues
missed
and
false
detections
encountered
in
small
object
detection
for
UAV
remote
sensing
images,
inadequacy
existing
algorithms
terms
complexity
generalization
ability,
we
propose
a
model
named
IA-YOLOv8
this
paper.
This
integrates
intra-group
multi-scale
fusion
attention
mechanism
adaptive
weighted
feature
approach.
extraction
phase,
employs
hybrid
pooling
strategy
that
combines
Avg
Max
to
replace
single
operation
used
original
SPPF
framework.
Such
modifications
enhance
model’s
ability
capture
minute
features
objects.
addition,
an
module
is
introduced,
which
capable
automatically
adjusting
weights
based
on
significance
contribution
at
different
scales
improve
sensitivity
Simultaneously,
lightweight
implemented,
aims
effectively
mitigate
background
interference
saliency
Experimental
results
indicate
proposed
has
parameter
quantity
10.9
MB,
attaining
average
precision
(mAP)
value
42.1%
Visdrone2019
test
set,
mAP
82.3%
DIOR
39.8%
AI-TOD
set.
All
these
outperform
algorithms,
demonstrating
superior
performance
task
sensing.
Язык: Английский
From Stationary to Nonstationary UAVs: Deep-Learning-Based Method for Vehicle Speed Estimation
Algorithms,
Год журнала:
2024,
Номер
17(12), С. 558 - 558
Опубликована: Дек. 6, 2024
The
development
of
smart
cities
relies
on
the
implementation
cutting-edge
technologies.
Unmanned
aerial
vehicles
(UAVs)
and
deep
learning
(DL)
models
are
examples
such
disruptive
technologies
with
diverse
industrial
applications
that
gaining
traction.
When
it
comes
to
road
traffic
monitoring
systems
(RTMs),
combination
UAVs
vision-based
methods
has
shown
great
potential.
Currently,
most
solutions
focus
analyzing
footage
captured
by
hovering
due
inherent
georeferencing
challenges
in
video
from
nonstationary
drones.
We
propose
an
innovative
method
capable
estimating
speed
using
both
stationary
UAVs.
process
involves
matching
each
pixel
input
frame
a
georeferenced
orthomosaic
feature-matching
algorithm.
Subsequently,
tracking-enabled
YOLOv8
object
detection
model
is
applied
detect
their
trajectories.
geographic
positions
these
moving
over
time
logged
JSON
format.
accuracy
this
was
validated
reference
measurements
recorded
laser
gun.
results
indicate
proposed
can
estimate
vehicle
speeds
absolute
error
as
low
0.53
km/h.
study
also
discusses
associated
problems
constraints
drone
proposes
strategies
for
minimizing
noise
inaccuracies.
Despite
challenges,
framework
demonstrates
considerable
potential
signifies
another
step
towards
automated
systems.
This
system
enables
transportation
modelers
realistically
capture
behavior
wider
area,
unlike
existing
roadside
camera
prone
blind
spots
limited
spatial
coverage.
Язык: Английский
Enhance the Concrete Crack Classification Based on a Novel Multi-Stage YOLOV10-ViT Framework
Sensors,
Год журнала:
2024,
Номер
24(24), С. 8095 - 8095
Опубликована: Дек. 19, 2024
Early
identification
of
concrete
cracks
and
multi-class
detection
can
help
to
avoid
future
deformation
or
collapse
in
structures.
Available
traditional
methodologies
require
enormous
effort
time.
To
overcome
such
difficulties,
current
vision-based
deep
learning
models
effectively
detect
classify
various
cracks.
This
study
introduces
a
novel
multi-stage
framework
for
crack
type
classification.
First,
the
recently
developed
YOLOV10
model
is
trained
possible
defective
regions
images.
After
that,
modified
vision
transformer
(ViT)
images
into
three
main
types:
normal,
simple
cracks,
multi-branched
The
evaluation
process
includes
feeding
test
model,
identifying
defect
regions,
finally
delivering
detected
ViT
which
decides
appropriate
those
regions.
Experiments
are
conducted
using
individual
proposed
framework.
improve
generation
ability,
multi-source
datasets
structures
used.
For
classification
part,
dataset
consisting
12,000
classes
utilized,
while
composed
materials
from
historical
buildings
containing
1116
with
their
corresponding
bounding
boxes,
utilized.
Results
prove
that
accurately
classifies
types
90.67%
precision,
90.03%
recall,
90.34%
F1-score.
results
also
show
outperforms
by
10.9%,
19.99%,
19.2%
F1-score,
respectively.
YOLOV10-ViT
be
integrated
construction
systems
based
on
obtain
early
warning
Язык: Английский