On the creativity of large language models
AI & Society,
Год журнала:
2024,
Номер
unknown
Опубликована: Ноя. 28, 2024
Abstract
Large
language
models
(LLMs)
are
revolutionizing
several
areas
of
Artificial
Intelligence.
One
the
most
remarkable
applications
is
creative
writing,
e.g.,
poetry
or
storytelling:
generated
outputs
often
astonishing
quality.
However,
a
natural
question
arises:
can
LLMs
be
really
considered
creative?
In
this
article,
we
first
analyze
development
under
lens
creativity
theories,
investigating
key
open
questions
and
challenges.
particular,
focus
our
discussion
on
dimensions
value,
novelty,
surprise
as
proposed
by
Margaret
Boden
in
her
work.
Then,
consider
different
classic
perspectives,
namely
product,
process,
press,
person.
We
discuss
set
“easy”
“hard”
problems
machine
creativity,
presenting
them
relation
to
LLMs.
Finally,
examine
societal
impact
these
technologies
with
particular
industries,
analyzing
opportunities
offered,
challenges
arising
from
them,
potential
associated
risks,
both
legal
ethical
points
view.
Язык: Английский
A Manager and an AI Walk into a Bar: Does ChatGPT Make Biased Decisions Like We Do?
SSRN Electronic Journal,
Год журнала:
2023,
Номер
unknown
Опубликована: Янв. 1, 2023
Large
language
models
(LLMs)
such
as
ChatGPT
have
garnered
global
attention
recently,
with
a
promise
to
disrupt
and
revolutionize
business
operations.
As
managers
rely
more
on
artificial
intelligence
(AI)
technology,
there
is
an
urgent
need
understand
whether
are
systematic
biases
in
AI
decision-making
since
they
trained
human
data
feedback,
both
may
be
highly
biased.
This
paper
tests
broad
range
of
behavioral
commonly
found
humans
that
especially
relevant
operations
management.
We
although
can
much
less
biased
accurate
than
problems
explicit
mathematical/probabilistic
natures,
it
also
exhibits
many
possess,
when
the
complicated,
ambiguous,
implicit.
It
suffer
from
conjunction
bias
probability
weighting.
Its
preference
influenced
by
framing,
salience
anticipated
regret,
choice
reference.
struggles
process
ambiguous
information
evaluates
risks
differently
humans.
produce
responses
similar
heuristics
employed
humans,
prone
confirmation
bias.
To
make
these
issues
worse,
overconfident.
Our
research
characterizes
ChatGPT's
behaviors
showcases
for
researchers
businesses
consider
potentialAI
developing
employing
Язык: Английский
Theory Is All You Need: AI, Human Cognition, and Decision Making
SSRN Electronic Journal,
Год журнала:
2024,
Номер
unknown
Опубликована: Янв. 1, 2024
Artificial
intelligence
(AI)
now
matches
or
outperforms
human
in
an
astonishing
array
of
games,
tests,
and
other
cognitive
tasks
that
involve
high-level
reasoning
thinking.
Many
scholars
argue
that—due
to
bias
bounded
rationality—humans
should
(or
will
soon)
be
replaced
by
AI
situations
involving
cognition
strategic
decision
making.
We
disagree.
In
this
paper
we
first
trace
the
historical
origins
idea
artificial
as
a
form
computation
information
processing.
highlight
problems
with
analogy
between
computers
minds
input-output
devices,
using
large
language
models
example.
Human
cognition—in
important
instances—is
better
conceptualized
theorizing
rather
than
data
processing,
prediction,
even
Bayesian
updating.
Our
argument,
when
it
comes
cognition,
is
AI's
data-based
prediction
different
from
theory-based
causal
logic.
introduce
belief-data
(a)symmetries
difference
use
"heavier-than-air
flight"
example
our
arguments.
Theories
provide
mechanism
for
identifying
new
evidence,
way
"intervening"
world,
experimenting,
problem
solving.
conclude
discussion
implications
arguments
making,
including
role
human-AI
hybrids
might
play
process.
Язык: Английский
(Ir)rationality and cognitive biases in large language models
Royal Society Open Science,
Год журнала:
2024,
Номер
11(6)
Опубликована: Июнь 1, 2024
Do
large
language
models
(LLMs)
display
rational
reasoning?
LLMs
have
been
shown
to
contain
human
biases
due
the
data
they
trained
on;
whether
this
is
reflected
in
reasoning
remains
less
clear.
In
paper,
we
answer
question
by
evaluating
seven
using
tasks
from
cognitive
psychology
literature.
We
find
that,
like
humans,
irrationality
these
tasks.
However,
way
displayed
does
not
reflect
that
humans.
When
incorrect
answers
are
given
tasks,
often
ways
differ
human-like
biases.
On
top
of
this,
reveal
an
additional
layer
significant
inconsistency
responses.
Aside
experimental
results,
paper
seeks
make
a
methodological
contribution
showing
how
can
assess
and
compare
different
capabilities
types
models,
case
with
respect
reasoning.
Язык: Английский
Anchoring Bias in Large Language Models: An Experimental Study
Опубликована: Янв. 28, 2025
Large
Language
Models
(LLMs)
like
GPT-4
and
Gemini
have
significantly
advanced
artificial
intelligence
by
enabling
machines
to
generate
comprehend
human-like
text.
Despite
their
impressive
capabilities,
LLMs
are
not
free
of
limitations.
They
shown
various
biases.
While
much
research
has
explored
demographic
biases,
the
cognitive
biases
in
been
equally
studied.
This
study
delves
into
anchoring
bias,
a
bias
where
initial
information
disproportionately
influences
judgment.
Utilizing
an
experimental
dataset,
we
examine
how
manifests
verify
effectiveness
mitigation
strategies.
Our
findings
highlight
sensitivity
LLM
responses
biased
hints.
At
same
time,
our
experiments
show
that,
mitigate
one
needs
collect
hints
from
comprehensive
angles
prevent
being
anchored
individual
pieces
information,
while
simple
algorithms
such
as
Chain-of-Thought,
Thoughts
Principles,
Ignoring
Anchor
Hints,
Reflection
sufficient.
Язык: Английский
Biases, evolutionary mismatch and the comparative analysis of human versus artificial cognition: a comment on Macmillan-Scott and Musolesi (2024)
Royal Society Open Science,
Год журнала:
2025,
Номер
12(2)
Опубликована: Фев. 1, 2025
Язык: Английский
Natural Language Processing and Large Language Models
Springer texts in education,
Год журнала:
2025,
Номер
unknown, С. 117 - 142
Опубликована: Янв. 1, 2025
Язык: Английский
theoraizer: AI-assisted Theory Construction
Опубликована: Авг. 2, 2024
The
Causal
Loop
Diagram
(CLD)
method
is
a
technique
for
theory
construction
in
which
domain
experts
collaborate
to
identify
causal
relationships
between
variables.
However,
CLD
labor-intensive,
and
the
input
required
from
grows
quadratically
with
number
of
variables
involved.
This
limits
small
graphs.
Large
Language
Models
(LLMs),
their
advanced
text
processing
capabilities
extensive
knowledge
base,
can
efficiently
generate
large
amounts
content,
offering
potential
overcome
these
limitations.
paper
presents
theoraizer,
an
R
package
Shiny
app
that
enhances
by
integrating
LLMs
as
digital
extension
expert
group.
Researchers
use
theoraizer
define
list
putative
variables,
after
it
queries
LLM
links
drastically
reduces
amount
work
arrive
at
candidate
provides
scientists
standardized,
multi-stage
framework
constructing
theories.
Язык: Английский
Wait, It’s All Token Noise? Always Has Been: Interpreting LLM Behavior Using Shapley Value
SSRN Electronic Journal,
Год журнала:
2024,
Номер
unknown
Опубликована: Янв. 1, 2024
The
emergence
of
large
language
models
(LLMs)
has
opened
up
exciting
possibilities
for
simulating
human
behavior
and
cognitive
processes,
with
potential
applications
in
various
domains,
including
marketing
research
consumer
analysis.
However,
the
validity
utilizing
LLMs
as
stand-ins
subjects
remains
uncertain
due
to
glaring
divergences
that
suggest
fundamentally
different
underlying
processes
at
play
sensitivity
LLM
responses
prompt
variations.
This
paper
presents
a
novel
approach
based
on
Shapley
values
from
cooperative
game
theory
interpret
quantify
relative
contribution
each
component
model's
output.
Through
two
applications—a
discrete
choice
experiment
an
investigation
biases—we
demonstrate
how
value
method
can
uncover
what
we
term
"token
noise"
effects,
phenomenon
where
decisions
are
disproportionately
influenced
by
tokens
providing
minimal
informative
content.
raises
concerns
about
robustness
generalizability
insights
obtained
context
simulation.
Our
model-agnostic
extends
its
utility
proprietary
LLMs,
valuable
tool
marketers
researchers
strategically
optimize
prompts
mitigate
apparent
biases.
findings
underscore
need
more
nuanced
understanding
factors
driving
before
relying
them
substitutes
settings.
We
emphasize
importance
reporting
results
conditioned
specific
templates
exercising
caution
when
drawing
parallels
between
LLMs.
Язык: Английский