Analysis of Language-Model-Powered Chatbots for Query Resolution in PDF-Based Automotive Manuals DOI Creative Commons
Thaís Medeiros, Morsinaldo Medeiros, Mariana Azevedo

и другие.

Vehicles, Год журнала: 2023, Номер 5(4), С. 1384 - 1399

Опубликована: Окт. 16, 2023

In the current scenario of fast technological advancement, increasingly characterized by widespread adoption Artificial Intelligence (AI)-driven tools, significance autonomous systems like chatbots has been highlighted. Such systems, which are proficient in addressing queries based on PDF files, hold potential to revolutionize customer support and post-sales services automotive sector, resulting time resource optimization. Within this scenario, work explores Large Language Models (LLMs) create AI-assisted tools for assuming three distinct methods comparative analysis. For them, broad assessment criteria considered order encompass response accuracy, cost, user experience. The achieved results demonstrate that choice most adequate method context hinges selected criteria, with different practical implications. Therefore, provides insights into effectiveness applicability industry, particularly when interfacing manuals, facilitating implementation productive generative AI strategies meet demands sector.

Язык: Английский

Dissociating language and thought in large language models DOI
Kyle Mahowald, Anna A. Ivanova, Idan Blank

и другие.

Trends in Cognitive Sciences, Год журнала: 2024, Номер 28(6), С. 517 - 540

Опубликована: Март 19, 2024

Язык: Английский

Процитировано

115

Language writ large: LLMs, ChatGPT, meaning, and understanding DOI Creative Commons
Stevan Harnad

Frontiers in Artificial Intelligence, Год журнала: 2025, Номер 7

Опубликована: Фев. 12, 2025

Apart from what (little) OpenAI may be concealing us, we all know (roughly) how Large Language Models (LLMs) such as ChatGPT work (their vast text databases, statistics, vector representations, and huge number of parameters, next-word training, etc.). However, none us can say (hand on heart) that are not surprised by has proved to able do with these resources. This even driven some conclude actually understands. It is true it But also understand do. I will suggest hunches about benign “biases”—convergent constraints emerge at the LLM scale helping so much better than would have expected. These biases inherent in nature language itself, scale, they closely linked lacks , which direct sensorimotor grounding connect its words their referents propositions meanings. convergent related (1) parasitism indirect verbal grounding, (2) circularity definition, (3) “mirroring” production comprehension, (4) iconicity (5) computational counterparts human “categorical perception” category learning neural nets, perhaps (6) a conjecture Chomsky laws thought. The exposition form dialogue ChatGPT-4.

Язык: Английский

Процитировано

2

GPT and CLT: The impact of ChatGPT's level of abstraction on consumer recommendations DOI Creative Commons
Samuel N. Kirshner

Journal of Retailing and Consumer Services, Год журнала: 2023, Номер 76, С. 103580 - 103580

Опубликована: Окт. 20, 2023

This study explores how ChatGPT interprets information through the lens of Construal Level Theory (CLT). The findings show that exhibits an abstraction bias, generating responses consistent with a high-level construal. bias results in prioritising construal features (e.g., desirability) over low-level feasibility) consumer evaluation scenarios. Thus, recommendations differ significantly from traditional based on human decision-making. Applying CLT concepts to large language models provides essential insights into behaviour may evolve increasing prevalence and capability AI offers many promising avenues for future research.

Язык: Английский

Процитировано

25

Still no lie detector for language models: probing empirical and conceptual roadblocks DOI
Benjamin A. Levinstein, Daniel A. Herrmann

Philosophical Studies, Год журнала: 2024, Номер unknown

Опубликована: Фев. 17, 2024

Язык: Английский

Процитировано

11

Do Language Models’ Words Refer? DOI Creative Commons
Matthew Mandelkern, Tal Linzen

Computational Linguistics, Год журнала: 2024, Номер unknown, С. 1 - 10

Опубликована: Июнь 10, 2024

Abstract What do language models (LMs) with language? They can produce sequences of (mostly) coherent strings closely resembling English. But those sentences mean something, or are LMs simply babbling in a convincing simulacrum use? We address one aspect this broad question: whether LMs’ words refer, that is, achieve “word-to-world” connections. There is prima facie reason to think they not, since not interact the world way ordinary users do. Drawing on externalist tradition philosophy language, we argue appearances misleading: Even if inputs text, text natural histories, and may suffice for refer.

Язык: Английский

Процитировано

10

Improving Generalization Beyond Training Data with Compositional Generalization in Large Language Models DOI Open Access

Wong Ho-tin,

Gar-lai Yip

Опубликована: Май 20, 2024

Enhancing compositional generalization in language models addresses a crucial challenge natural processing, significantly improving their ability to understand and generate novel combinations of known concepts. The investigation utilized the Mistral 7x8B model, employing advanced data augmentation refined training methodologies enhance performance. By incorporating diverse challenging compositions during training, model demonstrated substantial gains standard evaluation metrics, including accuracy, precision, recall, F1-score. Specialized metrics such as accuracy contextual coherence also showed marked improvement, reflecting model's enhanced capacity correct contextually relevant outputs when faced with compositions. study further highlighted significant reduction hallucination rates, underscoring increased logical consistency factual accuracy. This was statistically significant, indicating robust enhancement Qualitative analysis corroborated these findings, revealing more coherent narratives accurate information retrieval generated responses. These improvements are particularly important for real-world applications where reliability appropriateness essential. comprehensive effectiveness proposed techniques, providing valuable insights into underlying mechanisms that contribute improved findings underscore importance iterative experimentation validation refining architectures techniques. advancing capabilities models, this research contributes development robust, flexible, reliable AI systems capable handling broader range linguistic tasks greater understanding.

Язык: Английский

Процитировано

8

Synergistic Integration of Large Language Models and Cognitive Architectures for Robust AI: An Exploratory Analysis DOI Open Access
Oscar J. Romero, John Zimmerman, Aaron Steinfeld

и другие.

Proceedings of the AAAI Symposium Series, Год журнала: 2024, Номер 2(1), С. 396 - 405

Опубликована: Янв. 22, 2024

This paper explores the integration of two AI subdisciplines employed in development artificial agents that exhibit intelligent behavior: Large Language Models (LLMs) and Cognitive Architectures (CAs). We present three approaches, each grounded theoretical models supported by preliminary empirical evidence. The modular approach, which introduces four with varying degrees integration, makes use chain-of-thought prompting, draws inspiration from augmented LLMs, Common Model Cognition, simulation theory cognition. agency motivated Society Mind LIDA cognitive architecture, proposes formation agent collections interact at micro macro levels, driven either LLMs or symbolic components. neuro-symbolic takes CLARION a model where bottom-up learning extracts representations an LLM layer top-down guidance utilizes to direct prompt engineering layer. These approaches aim harness strengths both CAs, while mitigating their weaknesses, thereby advancing more robust systems. discuss tradeoffs challenges associated approach.

Язык: Английский

Процитировано

7

Can large language models help augment English psycholinguistic datasets? DOI Creative Commons
Sean Trott

Behavior Research Methods, Год журнала: 2024, Номер 56(6), С. 6082 - 6100

Опубликована: Янв. 23, 2024

Research on language and cognition relies extensively psycholinguistic datasets or "norms". These contain judgments of lexical properties like concreteness age acquisition, can be used to norm experimental stimuli, discover empirical relationships in the lexicon, stress-test computational models. However, collecting human at scale is both time-consuming expensive. This issue compounded for multi-dimensional norms those incorporating context. The current work asks whether large models (LLMs) leveraged augment creation large, English. I use GPT-4 collect multiple kinds semantic (e.g., word similarity, contextualized sensorimotor associations, iconicity) English words compare these against "gold standard". For each dataset, find that GPT-4's are positively correlated with judgments, some cases rivaling even exceeding average inter-annotator agreement displayed by humans. then identify several ways which LLM-generated differ from human-generated systematically. also perform "substitution analyses", demonstrate replacing a statistical model does not change sign parameter estimates (though select cases, there significant changes their magnitude). conclude discussing considerations limitations associated general, including concerns data contamination, choice LLM, external validity, construct quality. Additionally, all (over 30,000 total) made available online further analysis.

Язык: Английский

Процитировано

7

The new paradigm in machine learning – foundation models, large language models and beyond: a primer for physicians DOI Creative Commons
Ian Scott, Guido Zuccon

Internal Medicine Journal, Год журнала: 2024, Номер 54(5), С. 705 - 715

Опубликована: Май 1, 2024

Abstract Foundation machine learning models are deep capable of performing many different tasks using data modalities such as text, audio, images and video. They represent a major shift from traditional task‐specific prediction models. Large language (LLM), brought to wide public prominence in the form ChatGPT, text‐based foundational that have potential transform medicine by enabling automation range tasks, including writing discharge summaries, answering patients questions assisting clinical decision‐making. However, not without risk can potentially cause harm if their development, evaluation use devoid proper scrutiny. This narrative review describes types LLM, emerging applications limitations bias likely future translation into practice.

Язык: Английский

Процитировано

7

Nanjing Yunjin intelligent question-answering system based on knowledge graphs and retrieval augmented generation technology DOI Creative Commons
Liang Xu, Lu Lu,

Minglu Liu

и другие.

Heritage Science, Год журнала: 2024, Номер 12(1)

Опубликована: Апрель 9, 2024

Abstract Nanjing Yunjin, a traditional Chinese silk weaving craft, is celebrated globally for its unique local characteristics and exquisite workmanship, forming an integral part of the world's intangible cultural heritage. However, with advancement information technology, experiential knowledge Yunjin production process predominantly stored in text format. As highly specialized vertical domain, this not readily convert into usable data. Previous studies on graph-based Question-Answering System have partially addressed issue. graphs need to be constantly updated rely predefined entities relationship types. Faced ambiguous or complex natural language problems, graph retrieval faces some challenges. Therefore, study proposes that integrates Knowledge Graphs Retrieval Augmented Generation techniques. In system, ROBERTA model first utilized vectorize textual information, delving deep semantics unveil profound connotations. Additionally, FAISS vector database employed efficient storage achieving semantic match between questions answers. Ultimately, related results are fed Large Language Model enhanced generation, aiming more accurate generation outcomes improving interpretability logic System. This research merges technologies like embedding, vectorized retrieval, overcome limitations graphs-based terms updating, dependency types, understanding. implementation testing shown Intelligent System, constructed basis Generation, possesses broader base considers context, resolving issues polysemy, vague language, sentence ambiguity, efficiently accurately generates answers queries. significantly facilitates utilization knowledge, providing paradigm constructing other heritages, holds substantial theoretical practical significance exploration discovery structure human heritage, promoting inheritance protection.

Язык: Английский

Процитировано

6