Trends in Cognitive Sciences,
Journal Year:
2023,
Volume and Issue:
28(2), P. 97 - 112
Published: Nov. 15, 2023
Prominent
accounts
of
sentient
behavior
depict
brains
as
generative
models
organismic
interaction
with
the
world,
evincing
intriguing
similarities
current
advances
in
artificial
intelligence
(AI).
However,
because
they
contend
control
purposive,
life-sustaining
sensorimotor
interactions,
living
organisms
are
inextricably
anchored
to
body
and
world.
Unlike
passive
learned
by
AI
systems,
must
capture
sensory
consequences
action.
This
allows
embodied
agents
intervene
upon
their
worlds
ways
that
constantly
put
best
test,
thus
providing
a
solid
bedrock
is
–
we
argue
essential
development
genuine
understanding.
We
review
resulting
implications
consider
future
directions
for
AI.
Proceedings of the National Academy of Sciences,
Journal Year:
2024,
Volume and Issue:
121(10)
Published: Feb. 29, 2024
During
real-time
language
comprehension,
our
minds
rapidly
decode
complex
meanings
from
sequences
of
words.
The
difficulty
doing
so
is
known
to
be
related
words’
contextual
predictability,
but
what
cognitive
processes
do
these
predictability
effects
reflect?
In
one
view,
reflect
facilitation
due
anticipatory
processing
words
that
are
predictable
context.
This
view
predicts
a
linear
effect
on
demand.
another
the
costs
probabilistic
inference
over
sentence
interpretations.
either
logarithmic
or
superlogarithmic
demand,
depending
whether
it
assumes
pressures
toward
uniform
distribution
information
time.
empirical
record
currently
mixed.
Here,
we
revisit
this
question
at
scale:
We
analyze
six
reading
datasets,
estimate
next-word
probabilities
with
diverse
statistical
models,
and
model
times
using
recent
advances
in
nonlinear
regression.
Results
support
word
difficulty,
which
favors
as
key
component
human
processing.
Proceedings of the National Academy of Sciences,
Journal Year:
2024,
Volume and Issue:
121(45)
Published: Oct. 29, 2024
Eleven
large
language
models
(LLMs)
were
assessed
using
40
bespoke
false-belief
tasks,
considered
a
gold
standard
in
testing
theory
of
mind
(ToM)
humans.
Each
task
included
scenario,
three
closely
matched
true-belief
control
scenarios,
and
the
reversed
versions
all
four.
An
LLM
had
to
solve
eight
scenarios
single
task.
Older
solved
no
tasks;
Generative
Pre-trained
Transformer
(GPT)-3-davinci-003
(from
November
2022)
ChatGPT-3.5-turbo
March
2023)
20%
ChatGPT-4
June
75%
matching
performance
6-y-old
children
observed
past
studies.
We
explore
potential
interpretation
these
results,
including
intriguing
possibility
that
ToM-like
ability,
previously
unique
humans,
may
have
emerged
as
an
unintended
by-product
LLMs'
improving
skills.
Regardless
how
we
interpret
outcomes,
they
signify
advent
more
powerful
socially
skilled
AI-with
profound
positive
negative
implications.
Nature Communications,
Journal Year:
2024,
Volume and Issue:
15(1)
Published: March 30, 2024
Contextual
embeddings,
derived
from
deep
language
models
(DLMs),
provide
a
continuous
vectorial
representation
of
language.
This
embedding
space
differs
fundamentally
the
symbolic
representations
posited
by
traditional
psycholinguistics.
We
hypothesize
that
areas
in
human
brain,
similar
to
DLMs,
rely
on
represent
To
test
this
hypothesis,
we
densely
record
neural
activity
patterns
inferior
frontal
gyrus
(IFG)
three
participants
using
dense
intracranial
arrays
while
they
listened
30-minute
podcast.
From
these
fine-grained
spatiotemporal
recordings,
derive
for
each
word
(i.e.,
brain
embedding)
patient.
Using
stringent
zero-shot
mapping
demonstrate
embeddings
IFG
and
DLM
contextual
have
common
geometric
patterns.
The
allow
us
predict
given
left-out
based
solely
its
geometrical
relationship
other
non-overlapping
words
Furthermore,
show
capture
geometry
better
than
static
embeddings.
exposes
vector-based
code
natural
processing
brain.
Neuron,
Journal Year:
2024,
Volume and Issue:
112(5), P. 698 - 717
Published: Feb. 9, 2024
Large
language
models
(LLMs)
are
a
new
asset
class
in
the
machine-learning
landscape.
Here
we
offer
primer
on
defining
properties
of
these
modeling
techniques.
We
then
reflect
modes
investigation
which
LLMs
can
be
used
to
reframe
classic
neuroscience
questions
deliver
fresh
answers.
reason
that
have
potential
(1)
enrich
datasets
by
adding
valuable
meta-information,
such
as
advanced
text
sentiment,
(2)
summarize
vast
information
sources
overcome
divides
between
siloed
communities,
(3)
enable
previously
unthinkable
fusion
disparate
relevant
brain,
(4)
help
deconvolve
cognitive
concepts
most
usefully
grasp
phenomena
and
much
more.
Nature Communications,
Journal Year:
2025,
Volume and Issue:
16(1)
Published: Jan. 18, 2025
Large
Language
Models
(LLMs)
have
shown
success
in
predicting
neural
signals
associated
with
narrative
processing,
but
their
approach
to
integrating
context
over
large
timescales
differs
fundamentally
from
that
of
the
human
brain.
In
this
study,
we
show
how
brain,
unlike
LLMs
process
text
windows
parallel,
integrates
short-term
and
long-term
contextual
information
through
an
incremental
mechanism.
Using
fMRI
data
219
participants
listening
spoken
narratives,
first
demonstrate
predict
brain
activity
effectively
only
when
using
short
up
a
few
dozen
words.
Next,
introduce
alternative
LLM-based
incremental-context
model
combines
incoming
aggregated,
dynamically
updated
summary
prior
context.
This
significantly
enhances
prediction
higher-order
regions
involved
long-timescale
processing.
Our
findings
reveal
brain's
hierarchical
temporal
processing
mechanisms
enable
flexible
integration
time,
providing
valuable
insights
for
both
cognitive
neuroscience
AI
development.
Journal of Neuroscience,
Journal Year:
2022,
Volume and Issue:
42(39), P. 7412 - 7430
Published: Aug. 24, 2022
To
understand
language,
we
must
infer
structured
meanings
from
real-time
auditory
or
visual
signals.
Researchers
have
long
focused
on
word-by-word
structure
building
in
working
memory
as
a
mechanism
that
might
enable
this
feat.
However,
some
argued
language
processing
does
not
typically
involve
rich
building,
and/or
apparent
effects
are
underlyingly
driven
by
surprisal
(how
predictable
word
is
context).
Consistent
with
alternative,
recent
behavioral
studies
of
naturalistic
control
for
surprisal
shown
clear
effects.
In
fMRI
study,
investigate
range
theory-driven
predictors
demand
during
comprehension
humans
both
sexes
under
rigorous
controls.
addition,
address
related
debate
about
whether
the
mechanisms
involved
specialized
domain
general.
do
so,
each
participant,
functionally
localize
(1)
language-selective
network
and
(2)
“multiple-demand”
network,
which
supports
across
domains.
Results
show
robust
surprisal-independent
no
effect
multiple-demand
network.
Our
findings
thus
support
view
involves
computationally
demanding
operations
memory,
addition
to
any
prediction-related
mechanisms.
Further,
these
appear
be
primarily
conducted
same
neural
resources
store
linguistic
knowledge,
evidence
involvement
brain
regions
known
SIGNIFICANCE
STATEMENT
This
study
uses
signatures
(WM)
story
listening,
using
broad
theoretically
motivated
estimates
WM
demand.
strong
distinct
predictability.
demands
register
regions,
rather
than
previously
been
associated
nonlinguistic
core
role
incremental
processing,
language.