Frontiers in Psychology,
Год журнала:
2024,
Номер
15
Опубликована: Окт. 29, 2024
In
recent
years,
the
capabilities
of
Large
Language
Models
(LLMs),
such
as
ChatGPT,
to
imitate
human
behavioral
patterns
have
been
attracting
growing
interest
from
experimental
psychology.
Although
ChatGPT
can
successfully
generate
accurate
theoretical
and
inferential
information
in
several
fields,
its
ability
exhibit
a
Theory
Mind
(ToM)
is
topic
debate
literature.
Impairments
ToM
are
considered
responsible
for
social
difficulties
many
clinical
conditions,
Autism
Spectrum
Disorder
(ASD).
Some
studies
showed
that
pass
classical
tasks,
however,
response
style
used
by
LLMs
solve
advanced
comparing
their
abilities
with
those
typical
development
(TD)
individuals
populations,
has
not
explored.
this
preliminary
study,
we
administered
Advanced
Test
Emotion
Attribution
Task
3.5
ChatGPT-4
compared
responses
an
ASD
TD
group.
Our
results
two
had
higher
accuracy
understanding
mental
states,
although
ChatGPT-3.5
failed
more
complex
states.
emotional
performed
significantly
worse
than
TDs
but
did
differ
ASDs,
showing
difficulty
negative
emotions.
achieved
accuracy,
recognizing
sadness
anger
persisted.
The
adopted
both
appeared
verbose,
repetitive,
tending
violate
Grice's
maxims.
This
conversational
seems
similar
high-functioning
ASDs.
Clinical
implications
potential
applications
discussed.
Patterns,
Год журнала:
2025,
Номер
6(2), С. 101176 - 101176
Опубликована: Фев. 1, 2025
Large
language
models
(LLMs)
have
demonstrated
performance
approaching
human
levels
in
tasks
such
as
long-text
comprehension
and
mathematical
reasoning,
but
they
remain
black-box
systems.
Understanding
the
reasoning
bottlenecks
of
LLMs
remains
a
critical
challenge,
these
limitations
are
deeply
tied
to
their
internal
architecture.
Attention
heads
play
pivotal
role
thought
share
similarities
with
brain
functions.
In
this
review,
we
explore
roles
mechanisms
attention
help
demystify
processes
LLMs.
We
first
introduce
four-stage
framework
inspired
by
process.
Using
framework,
review
existing
research
identify
categorize
functions
specific
heads.
Additionally,
analyze
experimental
methodologies
used
discover
special
further
summarize
relevant
evaluation
methods
benchmarks.
Finally,
discuss
current
propose
several
potential
future
directions.
Royal Society Open Science,
Год журнала:
2025,
Номер
12(2)
Опубликована: Фев. 1, 2025
We
examine
whether
a
leading
AI
system,
GPT-4,
understands
text
as
well
humans
do,
first
using
well-established
standardized
test
of
discourse
comprehension.
On
this
test,
GPT-4
performs
slightly,
but
not
statistically
significantly,
better
than
given
the
very
high
level
human
performance.
Both
and
make
correct
inferences
about
information
that
is
explicitly
stated
in
text,
critical
understanding.
Next,
we
use
more
difficult
passages
to
determine
could
allow
larger
differences
between
humans.
does
considerably
on
do
school
university
students
for
whom
these
are
designed,
admission
tests
student
reading
Deeper
exploration
GPT-4's
performance
material
from
one
reveals
generally
accepted
signatures
genuine
understanding,
namely
generalization
inference.
Entertainment Computing,
Год журнала:
2024,
Номер
52, С. 100810 - 100810
Опубликована: Июль 4, 2024
This
article
presents
a
novel
and
highly
interactive
process
to
generate
natural
language
narratives
based
on
our
ongoing
work
semiotic
relations,
providing
four
criteria
for
composing
new
from
existing
stories.
The
wide
applicability
of
this
reconstruction
is
suggested
by
reputed
literary
scholar's
deconstructive
claim
that
can
often
be
shown
tissue
previous
narratives.
Along,
respectively,
three
axes
–
syntagmatic,
paradigmatic,
meronymic
stories
yield
the
combination,
imitation,
or
expansion
an
iconic
scene;
lastly,
story
may
emerge
through
reversal
via
antithetic
consideration,
i.e.,
adoption
opposite
values.
Targeting
casual
users,
we
present
fully
operational
prototype
with
simple
user-friendly
interface
incorporates
AI
agent,
namely
ChatGPT.
prototype,
in
coauthor
capacity,
generates
context-compatible
sequences
events
storyboard
format
using
backward-chaining
abductive
reasoning
(employing
Stable
Diffusion
draw
scene
illustrations),
conforming
as
much
possible
user's
authorial
instructions.
extensive
repertoire
book
movie
summaries
available
agent
obviates
need
manually
supply
laborious
error-prone
context
specifications.
A
user
study
was
conducted
evaluate
experience
satisfaction
generated
preliminary
findings
suggest
approach
has
potential
enhance
quality
while
offering
positive
experience.
Proceedings of the AAAI Symposium Series,
Год журнала:
2024,
Номер
3(1), С. 115 - 124
Опубликована: Май 20, 2024
Domain-specific
knowledge
can
significantly
contribute
to
addressing
a
wide
variety
of
vision
tasks.
However,
the
generation
such
entails
considerable
human
labor
and
time
costs.
This
study
investigates
potential
Large
Language
Models
(LLMs)
in
generating
providing
domain-specific
information
through
semantic
embeddings.
To
achieve
this,
an
LLM
is
integrated
into
pipeline
that
utilizes
Knowledge
Graphs
pre-trained
vectors
context
Vision-based
Zero-shot
Object
State
Classification
task.
We
thoroughly
examine
behavior
extensive
ablation
study.
Our
findings
reveal
integration
LLM-based
embeddings,
combination
with
general-purpose
leads
substantial
performance
improvements.
Drawing
insights
from
this
study,
we
conduct
comparative
analysis
against
competing
models,
thereby
highlighting
state-of-the-art
achieved
by
proposed
approach.
Organizational Research Methods,
Год журнала:
2024,
Номер
unknown
Опубликована: Авг. 28, 2024
When
assessing
text,
supervised
natural
language
processing
(NLP)
models
have
traditionally
been
used
to
measure
targeted
constructs
in
the
organizational
sciences.
However,
these
require
significant
resources
develop.
Emerging
“off-the-shelf”
large
(LLM)
offer
a
way
evaluate
without
building
customized
models.
it
is
unclear
whether
off-the-shelf
LLMs
accurately
score
and
what
evidence
necessary
infer
validity.
In
this
study,
we
compared
validity
of
NLP
LLM
(ChatGPT-3.5
ChatGPT-4).
Across
six
datasets
thousands
comments,
found
that
produced
scores
were
more
reliable
than
human
coders.
even
though
not
specifically
developed
for
purpose,
produce
similar
psychometric
properties
as
models,
with
slightly
less
favorable
properties.
We
connect
findings
broader
validation
considerations
present
decision
chart
guide
researchers
practitioners
on
how
they
can
use
constructs,
including
guidance
be
“transported”
new
contexts.
European Journal of Futures Research,
Год журнала:
2024,
Номер
12(1)
Опубликована: Авг. 24, 2024
Abstract
This
paper
explores
the
potential
of
a
multidisciplinary
approach
to
testing
and
aligning
artificial
intelligence
(AI),
specifically
focusing
on
large
language
models
(LLMs).
Due
rapid
development
wide
application
LLMs,
challenges
such
as
ethical
alignment,
controllability,
predictability
these
emerged
global
risks.
study
investigates
an
innovative
simulation-based
multi-agent
system
within
virtual
reality
framework
that
replicates
real-world
environment.
The
is
populated
by
automated
'digital
citizens,'
simulating
complex
social
structures
interactions
examine
optimize
AI.
Application
various
theories
from
fields
sociology,
psychology,
computer
science,
physics,
biology,
economics
demonstrates
possibility
more
human-aligned
socially
responsible
purpose
digital
environment
provide
dynamic
platform
where
advanced
AI
agents
can
interact
make
independent
decisions,
thereby
mimicking
realistic
scenarios.
actors
in
this
city,
operated
serve
primary
agents,
exhibiting
high
degrees
autonomy.
While
shows
immense
potential,
there
are
notable
limitations,
most
significantly
unpredictable
nature
dynamics.
research
endeavors
contribute
refinement
AI,
emphasizing
integration
social,
ethical,
theoretical
dimensions
for
future
research.
SSRN Electronic Journal,
Год журнала:
2024,
Номер
unknown
Опубликована: Янв. 1, 2024
This
paper
delves
into
behavioral
biases
in
the
economic
decision-making
of
LLMs
across
English,
Chinese,
Spanish,
and
French,
focusing
specifically
on
mental
accounting.
Our
investigation
comprises
three
distinct
components.
First,
we
analyze
LLMs'
arithmetic
using
prospect
theory,
revealing
that
Spanish
French
exhibit
human-like
value
functions,
while
English
Chinese
lack
loss
aversion.
Additionally,
explore
application
hedonic
framing,
noting
their
alignment
with
human
single-type
outcomes
but
divergence
mixed
loss-gain
scenarios.
The
second
component
evaluates
influence
accounting
financial
decision-making,
particularly
contexts
where
consumption
is
either
concurrent
or
separate
from
transactions.
findings
reveal
only
derive
utility
consistently
demonstrate
a
ability
to
distinguish
between
items
intended
for
immediate
use
those
future
consumption.
In
final
component,
examine
dynamic
tend
temporally
segregate
losses,
prefer
separating
gains,
deviating
humans
segregating
both.
indicate
mimic
certain
aspects,
significant
differences
persist.
These
insights
underscore
need
caution
when
employing
understand
consumer
preferences
simulate
decision-making.
East African Journal of Information Technology,
Год журнала:
2024,
Номер
7(1), С. 188 - 201
Опубликована: Авг. 15, 2024
Recent
advancements
in
Artificial
Intelligence
(AI),
particularly
the
advanced
machine
learning
for
Natural
Language
Processing
(NLP)
paradigm,
have
led
to
development
of
powerful
Large
Models
(LLMs)
capable
impressive
feats
tasks
like
translation,
text
summarisation,
generation
and
code
generation.
However,
a
critical
challenge
hindering
their
real-world
deployment
is
susceptibility
hallucinations,
where
they
generate
plausible
looking
but
factually
incorrect
outputs.
These
limitations
come
with
adverse
effects,
such
as
propagation
misinformation
reducing
user
trustworthiness
related
technologies,
even
when
possess
transformative
potential
various
sectors.
This
study
aims
enhance
performance
LLMs
by
presenting
new
strategy
that
combines
grammar-aware
prompt
engineering
(GAPE)
formal
methods
(FMs)
leverage
synergy
LLM
process
logic.
We
argue
combining
linguistic
principles
using
GAPE
constructing
basis
structures
FMs,
we
could
improve
LLM's
ability
analyse
language,
decrease
ambiguity
prompts,
consistency
output,
eventually,
greatly
diminish
hallucinations.
To
do
this,
propose
collaboration
between
linguists
AI
experts
while
also
providing
specialised
training
emphasises
precision.
Additionally,
suggest
implementing
iterative
design
procedures
use
FM
continuously
LLMs.
By
following
these
techniques,
may
create
future
which
are
more
trustworthy
wide
range
users
cases
reliable
technologies
efficient
practical
situations