arXiv (Cornell University),
Journal Year:
2023,
Volume and Issue:
unknown
Published: Jan. 1, 2023
Large
language
models
(LLMs)
have
been
successfully
applied
to
software
engineering
tasks,
including
program
repair.
However,
their
application
in
search-based
techniques
such
as
Genetic
Improvement
(GI)
is
still
largely
unexplored.
In
this
paper,
we
evaluate
the
use
of
LLMs
mutation
operators
for
GI
improve
search
process.
We
expand
Gin
Java
toolkit
call
OpenAI's
API
generate
edits
JCodec
tool.
randomly
sample
space
using
5
different
edit
types.
find
that
number
patches
passing
unit
tests
up
75%
higher
with
LLM-based
than
standard
Insert
edits.
Further,
observe
found
are
generally
less
diverse
compared
ran
local
runtime
improvements.
Although
many
improving
by
LLM-enhanced
GI,
best
patch
was
GI.
Education Sciences,
Journal Year:
2024,
Volume and Issue:
14(6), P. 656 - 656
Published: June 17, 2024
This
study
addresses
the
significant
challenge
posed
by
use
of
Large
Language
Models
(LLMs)
such
as
ChatGPT
on
integrity
online
examinations,
focusing
how
these
models
can
undermine
academic
honesty
demonstrating
their
latent
and
advanced
reasoning
capabilities.
An
iterative
self-reflective
strategy
was
developed
for
invoking
critical
thinking
higher-order
in
LLMs
when
responding
to
complex
multimodal
exam
questions
involving
both
visual
textual
data.
The
proposed
demonstrated
evaluated
real
subject
experts
performance
(GPT-4)
with
vision
estimated
an
additional
dataset
600
text
descriptions
questions.
results
indicate
that
invoke
multi-hop
capabilities
within
LLMs,
effectively
steering
them
towards
correct
answers
integrating
from
each
modality
into
final
response.
Meanwhile,
considerable
proficiency
being
able
answer
across
12
subjects.
These
findings
prior
assertions
about
limitations
emphasise
need
robust
security
measures
proctoring
systems
more
sophisticated
mitigate
potential
misconduct
enabled
AI
technologies.
This
study
presents
a
novel
approach
to
enhance
Large
Language
Models
(LLMs)
like
Alpaca
by
dynamically
integrating
real-time
information.
method
addresses
the
issue
of
content
hallucination
and
data
relevancy
automatically
collecting
current
from
credible
sources
into
model
prompts.
Experiments
show
significant
improvement
in
accuracy
decrease
hallucination,
with
manageable
increase
response
time.
The
research
underscores
potential
integration
making
LLMs
more
accurate
contextually
relevant,
setting
foundation
for
future
advancements
dynamic
processing
AI.
Journal of the Academy of Marketing Science,
Journal Year:
2024,
Volume and Issue:
unknown
Published: Aug. 26, 2024
Abstract
Nowadays,
we
are
witnessing
the
exponential
growth
of
Generative
AI
(GenAI),
a
group
models
designed
to
produce
new
content.
This
technology
is
poised
revolutionize
marketing
research
and
practice.
Since
literature
about
GenAI
still
in
its
infancy,
offer
technical
overview
how
trained
they
Following
this,
construct
roadmap
for
future
on
marketing,
divided
into
two
main
domains.
The
first
domain
focuses
firms
can
harness
potential
throughout
innovation
process.
We
begin
by
discussing
changes
consumer
behavior
propose
questions
at
level.
then
connect
these
emerging
insights
with
corresponding
firm
strategies,
presenting
second
set
examines
likely
consequences
using
analyze:
(1)
relationship
between
market-based
assets
value,
(2)
skills,
preferences,
role
processes.
Applied Sciences,
Journal Year:
2024,
Volume and Issue:
14(7), P. 3036 - 3036
Published: April 4, 2024
ChatGPT
plays
significant
roles
in
the
third
decade
of
21st
Century.
Smart
cities
applications
can
be
integrated
with
various
fields.
This
research
proposes
an
approach
for
developing
large
language
models
using
generative
artificial
intelligence
suitable
small-
and
medium-sized
enterprises
limited
hardware
resources.
There
are
many
AI
systems
operation
development.
However,
technological,
human,
financial
resources
required
to
develop
impractical
enterprises.
In
this
study,
we
present
a
proposed
reduce
training
time
computational
cost
that
is
designed
automate
question–response
interactions
specific
domains
smart
cities.
The
model
utilises
BLOOM
as
its
backbone
maximum
effectiveness
We
have
conducted
set
experiments
on
several
datasets
associated
validate
model.
Experiments
English
Vietnamese
languages
been
combined
low-rank
adaptation
cost.
comparative
experimental
testing,
outperformed
‘Phoenix’
multilingual
chatbot
by
achieving
92%
performance
compared
‘ChatGPT’
benchmark.
Royal Society Open Science,
Journal Year:
2024,
Volume and Issue:
11(6)
Published: June 1, 2024
Do
large
language
models
(LLMs)
display
rational
reasoning?
LLMs
have
been
shown
to
contain
human
biases
due
the
data
they
trained
on;
whether
this
is
reflected
in
reasoning
remains
less
clear.
In
paper,
we
answer
question
by
evaluating
seven
using
tasks
from
cognitive
psychology
literature.
We
find
that,
like
humans,
irrationality
these
tasks.
However,
way
displayed
does
not
reflect
that
humans.
When
incorrect
answers
are
given
tasks,
often
ways
differ
human-like
biases.
On
top
of
this,
reveal
an
additional
layer
significant
inconsistency
responses.
Aside
experimental
results,
paper
seeks
make
a
methodological
contribution
showing
how
can
assess
and
compare
different
capabilities
types
models,
case
with
respect
reasoning.
Frontiers in Artificial Intelligence,
Journal Year:
2025,
Volume and Issue:
8
Published: April 1, 2025
This
article
provides
an
epistemological
analysis
of
current
attempts
explaining
how
the
relatively
simple
algorithmic
components
neural
language
models
(NLMs)
provide
them
with
genuine
linguistic
competence.
After
introducing
Transformer
architecture,
at
basis
most
NLMs,
paper
firstly
emphasizes
central
question
in
philosophy
AI
has
been
shifted
from
"can
machines
think?",
as
originally
put
by
Alan
Turing,
to
"how
can
pointing
explanatory
gap
for
NLMs.
Subsequently,
existing
strategies
functioning
NLMs
are
analyzed
argue
that
they,
however
debated,
do
not
differ
used
cognitive
science
explain
intelligent
behaviors
humans.
In
particular,
available
experimental
studies
turned
test
theory
mind,
discourse
entity
tracking,
and
property
induction
examined
under
light
functional
science;
so-called
copying
algorithm
head
phenomenon
a
shown
mechanist
explanation
in-context
learning;
finally,
pioneering
use
predict
brain
activation
patterns
when
processing
here
involve
what
we
call
co-simulation,
which
NLM
simulate
understand
each
other.
Frontiers in Psychology,
Journal Year:
2025,
Volume and Issue:
16
Published: April 1, 2025
Neural
language
models,
although
at
first
approximation
they
may
be
simply
described
as
predictors
of
the
next
token
in
a
given
sequence,
surprisingly
exhibit
linguistic
behaviors
akin
to
human
ones.
This
suggests
existence
an
underlying
sophisticated
cognitive
system
production.
intriguing
circumstance
has
inspired
adoption
psychological
theories
investigative
tools
and
present
research
falls
within
this
line
inquiry.
What
we
aim
establish
is
potential
core
coherent
integration
production,
metaphorically
parallel
speaker's
personal
identity.
To
investigate
this,
employed
well-established
theory
on
narrative
coherence
autobiographical
stories.
offers
theoretical
advantage
strong
correlation
between
high
integrative
level
knowledge
system.
It
also
provides
empirical
methodologies
for
quantifying
its
characteristic
dimensions
through
analysis
texts.
The
same
methodology
was
applied
2010
stories
generated
by
GPT-3.5
equal
number
from
GPT-4,
elicited
asking
models
assume
roles
that
included
variety
variables
such
gender,
mood,
age.
large
ensures
adequate
sampling
stochastic
nature
made
possible
thanks
automated
evaluation
procedure.
We
initially
asked
generate
192
stories,
which
were
then
analyzed
team
professional
psychologists.
Based
sample,
constructed
training
set
fine-tuning
automatic
evaluator.
Our
results
4020
overall
show
fully
with
data
subjects,
slightly
higher
values
case
GPT-4.
These
suggest
unification
comparable
self
beings.