ISPRS Journal of Photogrammetry and Remote Sensing,
Journal Year:
2023,
Volume and Issue:
203, P. 373 - 391
Published: Aug. 31, 2023
3D
point
cloud
panoptic
segmentation
is
the
combined
task
to
(i)
assign
each
a
semantic
class
and
(ii)
separate
points
in
into
object
instances.
Recently
there
has
been
an
increased
interest
such
comprehensive
scene
understanding,
building
on
rapid
advances
of
due
advent
deep
neural
networks.
Yet,
date
very
little
work
about
outdoor
mobile-mapping
data,
no
systematic
comparisons.
The
present
paper
tries
close
that
gap.
It
reviews
blocks
needed
assemble
pipeline
related
literature.
Moreover,
modular
set
up
perform
comprehensive,
experiments
assess
state
context
street
mapping.
As
byproduct,
we
also
provide
first
public
dataset
for
task,
by
extending
NPM3D
include
instance
labels.
That
our
source
code
are
publicly
available.1We
discuss
which
adaptations
need
adapt
current
methods
scenes
large
objects.
Our
study
finds
mobile
mapping
KPConv
performs
best
but
slower,
while
PointNet++
fastest
significantly
worse.
Sparse
CNNs
between.
Regardless
backbone,
clustering
embedding
features
better
than
using
shifted
coordinates.
Sensors,
Journal Year:
2022,
Volume and Issue:
22(24), P. 9577 - 9577
Published: Dec. 7, 2022
LiDAR
is
a
commonly
used
sensor
for
autonomous
driving
to
make
accurate,
robust,
and
fast
decision-making
when
driving.
The
in
the
perception
system,
especially
object
detection,
understand
environment.
Although
2D
detection
has
succeeded
during
deep-learning
era,
lack
of
depth
information
limits
understanding
environment
location.
Three-dimensional
sensors,
such
as
LiDAR,
give
3D
about
surrounding
environment,
which
essential
system.
Despite
attention
computer
vision
community
due
multiple
applications
robotics
driving,
there
are
challenges,
scale
change,
sparsity,
uneven
distribution
data,
occlusions.
Different
representations
data
methods
minimize
effect
sparsity
have
been
proposed.
This
survey
presents
LiDAR-based
feature-extraction
techniques
data.
coordinate
systems
differ
camera
datasets
methods.
Therefore,
summarized.
Then,
state-of-the-art
object-detection
reviewed
with
selected
comparison
among
Sensors,
Journal Year:
2022,
Volume and Issue:
22(21), P. 8268 - 8268
Published: Oct. 28, 2022
Fish
species
recognition
is
crucial
to
identifying
the
abundance
of
fish
in
a
specific
area,
controlling
production
management,
and
monitoring
ecosystem,
especially
endangered
species,
which
makes
accurate
essential.
In
this
work,
problem
formulated
as
an
object
detection
model
handle
multiple
single
image,
challenging
classify
using
simple
classification
network.
The
proposed
consists
MobileNetv3-large
VGG16
backbone
networks
SSD
head.
Moreover,
class-aware
loss
function
solve
class
imbalance
our
dataset.
takes
number
instances
each
into
account
gives
more
weight
those
with
smaller
instances.
This
can
be
applied
any
or
task
imbalanced
experimental
result
on
large-scale
reef
dataset,
SEAMAPD21,
shows
that
improves
over
original
by
up
79.7%.
Pascal
VOC
dataset
also
outperforms
model.
IEEE Sensors Journal,
Journal Year:
2023,
Volume and Issue:
23(4), P. 3378 - 3394
Published: Jan. 13, 2023
An
accurate
and
robust
perception
system
is
key
to
understanding
the
driving
environment
of
autonomous
robots.
Autonomous
needs
3-D
information
about
objects,
including
object’s
location
pose,
understand
clearly.
A
camera
sensor
widely
used
in
because
its
richness
color
texture,
low
price.
The
major
problem
with
lack
information,
which
necessary
environment.
In
addition,
scale
change
occlusion
make
object
detection
more
challenging.
Many
deep
learning-based
methods,
such
as
depth
estimation,
have
been
developed
solve
information.
This
survey
presents
image
bounding
box
encoding
techniques
evaluation
metrics.
image-based
methods
are
categorized
based
on
technique
estimate
an
image’s
insights
added
each
method.
Then,
state-of-the-art
(SOTA)
monocular
stereo
camera-based
summarized.
We
also
compare
performance
selected
models
present
challenges
future
directions
detection.
Deleted Journal,
Journal Year:
2024,
Volume and Issue:
37(4), P. 1625 - 1641
Published: March 11, 2024
Lung
diseases
represent
a
significant
global
health
threat,
impacting
both
well-being
and
mortality
rates.
Diagnostic
procedures
such
as
Computed
Tomography
(CT)
scans
X-ray
imaging
play
pivotal
role
in
identifying
these
conditions.
X-rays,
due
to
their
easy
accessibility
affordability,
serve
convenient
cost-effective
option
for
diagnosing
lung
diseases.
Our
proposed
method
utilized
the
Contrast-Limited
Adaptive
Histogram
Equalization
(CLAHE)
enhancement
technique
on
images
highlight
key
feature
maps
related
using
DenseNet201.
We
have
augmented
existing
Densenet201
model
with
hybrid
pooling
channel
attention
mechanism.
The
experimental
results
demonstrate
superiority
of
our
over
well-known
pre-trained
models,
VGG16,
VGG19,
InceptionV3,
Xception,
ResNet50,
ResNet152,
ResNet50V2,
ResNet152V2,
MobileNetV2,
DenseNet121,
DenseNet169,
achieves
impressive
accuracy,
precision,
recall,
F1-scores
95.34%,
97%,
96%,
respectively.
also
provide
visual
insights
into
model's
decision-making
process
Gradient-weighted
Class
Activation
Mapping
(Grad-CAM)
identify
normal,
pneumothorax,
atelectasis
cases.
terms
heatmap
may
help
radiologists
improve
diagnostic
abilities
labelling
processes.
Sensors,
Journal Year:
2022,
Volume and Issue:
22(21), P. 8463 - 8463
Published: Nov. 3, 2022
When
it
comes
to
some
essential
abilities
of
autonomous
ground
vehicles
(AGV),
detection
is
one
them.
In
order
safely
navigate
through
any
known
or
unknown
environment,
AGV
must
be
able
detect
important
elements
on
the
path.
Detection
applicable
both
on-road
and
off-road,
but
they
are
much
different
in
each
environment.
The
key
environment
that
identify
drivable
pathway
whether
there
obstacles
around
it.
Many
works
have
been
published
focusing
components
various
ways.
this
paper,
a
survey
most
recent
advancements
methods
intended
specifically
for
off-road
has
presented.
For
this,
we
divided
literature
into
three
major
groups:
positive
negative
obstacles.
Each
portion
further
multiple
categories
based
technology
used,
example,
single
sensor-based,
how
data
analyzed.
Furthermore,
added
critical
findings
technology,
challenges
associated
with
possible
future
directions.
Authors
believe
work
will
help
reader
finding
who
doing
similar
works.
Sensors,
Journal Year:
2022,
Volume and Issue:
22(18), P. 7010 - 7010
Published: Sept. 16, 2022
Three-dimensional
object
detection
is
crucial
for
autonomous
driving
to
understand
the
environment.
Since
pooling
operation
causes
information
loss
in
standard
CNN,
we
designed
a
wavelet-multiresolution-analysis-based
3D
network
without
operation.
Additionally,
instead
of
using
single
filter
like
convolution,
used
lower-frequency
and
higher-frequency
coefficients
as
filter.
These
filters
capture
more
relevant
parts
than
filter,
enlarging
receptive
field.
The
model
comprises
discrete
wavelet
transform
(DWT)
an
inverse
(IWT)
with
skip
connections
encourage
feature
reuse
contrasting
expanding
layers.
IWT
enriches
representation
by
fully
recovering
lost
details
during
downsampling
Element-wise
summation
was
decrease
computational
burden.
We
trained
Haar
Daubechies
(Db4)
wavelets.
two-level
decomposition
result
shows
that
can
build
lightweight
losing
significant
performance.
experimental
results
on
KITTI's
BEV
evaluation
benchmark
show
our
outperforms
PointPillars-based
up
14%
while
reducing
number
trainable
parameters.
Sensors,
Journal Year:
2023,
Volume and Issue:
23(8), P. 4024 - 4024
Published: April 16, 2023
Self-driving
vehicles
must
be
controlled
by
navigation
algorithms
that
ensure
safe
driving
for
passengers,
pedestrians
and
other
vehicle
drivers.
One
of
the
key
factors
to
achieve
this
goal
is
availability
effective
multi-object
detection
tracking
algorithms,
which
allow
estimate
position,
orientation
speed
on
road.
The
experimental
analyses
conducted
so
far
have
not
thoroughly
evaluated
effectiveness
these
methods
in
road
scenarios.
To
aim,
we
propose
paper
a
benchmark
modern
applied
image
sequences
acquired
camera
installed
board
vehicle,
namely,
videos
available
BDD100K
dataset.
proposed
framework
allows
evaluate
22
different
combinations
using
metrics
highlight
positive
contribution
limitations
each
module
considered
algorithms.
analysis
results
points
out
best
method
currently
combination
ConvNext
QDTrack,
but
also
images
substantially
improved.
Thanks
our
analysis,
conclude
evaluation
should
extended
considering
specific
aspects
autonomous
scenarios,
such
as
multi-class
problem
formulation
distance
from
targets,
simulating
impact
errors
safety.
The Visual Computer,
Journal Year:
2024,
Volume and Issue:
unknown
Published: Jan. 29, 2024
Abstract
Point
clouds
consist
of
3D
data
points
and
are
among
the
most
considerable
formats
for
representations.
Their
popularity
is
due
to
their
broad
application
areas,
such
as
robotics
autonomous
driving,
employment
in
basic
vision
tasks
segmentation,
classification,
detection.
However,
processing
point
challenging
compared
other
visual
forms
images,
mainly
unstructured
nature.
Deep
learning
(DL)
has
been
established
a
powerful
tool
processing,
reporting
remarkable
performance
enhancements
traditional
methods
all
2D
tasks.
However
new
challenges
emerging
when
it
comes
clouds.
This
work
aims
guide
future
research
by
providing
systematic
review
DL
on
clouds,
holistically
covering
technologies
cloud
formation
reviewed
each
other.
The
discussed,
state-of-the-art
models’
performances
focusing
solutions.
Moreover,
this
popular
benchmark
datasets
summarized
based
task-oriented
applications,
aiming
highlight
existing
constraints
comparatively
evaluate
them.
Future
directions
upcoming
trends
also
highlighted.
2021 International Conference on 3D Vision (3DV),
Journal Year:
2024,
Volume and Issue:
unknown, P. 179 - 189
Published: March 18, 2024
We
introduce
a
highly
efficient
method
for
panoptic
segmentation
of
large
3D
point
clouds
by
redefining
this
task
as
scalable
graph
clustering
problem.
This
approach
can
be
trained
using
only
local
auxiliary
tasks,
thereby
eliminating
the
resource-intensive
instance-matching
step
during
training.
Moreover,
our
formulation
easily
adapted
to
superpoint
paradigm,
further
increasing
its
efficiency.
allows
model
process
scenes
with
millions
points
and
thousands
objects
in
single
inference.
Our
method,
called
SuperCluster,
achieves
new
state-of-the-art
performance
two
indoor
scanning
datasets:
50.1
PQ
(+7.8)
S3DIS
Area
5,
58.7
(+25.2)
ScanNetV2.
also
set
first
large-scale
mobile
mapping
benchmarks:
KITTI-360
DALES.
With
209k
parameters,
is
over
30
times
smaller
than
best-competing
trains
up
15
faster.
code
pretrained
models
are
available
at
https://github.com/drprojects/superpoint_transformer.