Regulating Generative AI: Ethical Considerations and Explainability Benchmarks DOI Open Access
C.K. Luk,

Hoi-Lam Chung,

Wai-Kuen Yim

и другие.

Опубликована: Март 20, 2024

This study looks into the critical discussion surrounding ethical regulation and explainability of generative artificial intelligence (AI). Amidst rapid advancement AI technologies, this paper identifies explores multifaceted concerns that arise, highlighting paramount importance transparency, accountability, fairness. Through an examination existing regulatory frameworks introduction novel benchmarks for explainability, advocates a balanced approach fosters innovation while ensuring oversight. Case studies illustrate dual potential to benefit society pose significant challenges, underscoring complexity its integration various domains. The findings emphasize necessity dynamic mechanisms, interdisciplinary collaboration, ongoing research navigate landscape AI, aiming harness capabilities responsibly betterment humanity.

Язык: Английский

Dynamic Content Generation in Large Language Models with Real-Time Constraints DOI Open Access
Jianhua Hu,

Huixiang Gao,

Qing Yuan

и другие.

Опубликована: Апрель 29, 2024

The rapid evolution of natural language processing capabilities, driven by advancements in large models (LLMs), has opened new avenues for real-time interactive applications. However, the static nature conventional LLMs poses significant limitations when adapting to dynamic user inputs real time. Dynamic Content Generation System (DCGS) proposed this study addresses these challenges integrating a modular overlay system that enhances flexibility and responsiveness existing LLMs, such as GPT-2, without altering their core architecture. Through series controlled experiments involving diverse scenarios, system's performance was rigorously evaluated based on metrics response time, content accuracy, satisfaction. Results demonstrated DCGS could significantly decrease times while maintaining high levels accuracy satisfaction, underlining its potential applications requiring immediate generation tailored specifications. implementation highlights capacity support adaptation various applications, from live digital interactions personalized creation media outlets. not only engagement providing more swiftly but also offers scalable solution adaptable future AI technologies.

Язык: Английский

Процитировано

10

Benchmarking the Hallucination Tendency of Google Gemini and Moonshot Kimi DOI Open Access

Ruoxi Shan,

Qiang Ming,

Guang Hong

и другие.

Опубликована: Май 22, 2024

To evaluate the hallucination tendencies of state-of-the-art language models is crucial for improving their reliability and applicability across various domains. This article presents a comprehensive evaluation Google Gemini Kimi using HaluEval benchmark, focusing on key performance metrics such as accuracy, relevance, coherence, rate. demonstrated superior performance, particularly in maintaining low rates high contextual while Kimi, though robust, showed areas needing further refinement. The study highlights importance advanced training techniques optimization enhancing model efficiency accuracy. Practical recommendations future development are provided, emphasizing need continuous improvement rigorous to achieve reliable efficient models.

Язык: Английский

Процитировано

9

Investigating Hallucination Tendencies of Large Language Models in Japanese and English DOI Creative Commons

Hiromi Tsuruta,

Rio Sakaguchi

Research Square (Research Square), Год журнала: 2024, Номер unknown

Опубликована: Июнь 4, 2024

Abstract The increasing reliance on artificial intelligence for natural language processing has brought to light the issue of hallucinations in models, where models generate content that appears plausible but is factually incorrect. Exploring comparative hallucination tendencies Japanese and English reveals significant differences, highlighting importance understanding language-specific challenges model performance. A rigorous methodology was employed quantify frequency severity hallucinations, with comprehensive data collection from diverse sources both languages. Quantitative analysis indicated a higher propensity responses, attributed complex syntactical contextual structures language. Qualitative examples provided concrete illustrations errors encountered, demonstrating impact linguistic cultural factors. findings emphasize necessity more linguistically contextually rich training datasets, along advanced fact-checking mechanisms, improve reliability models. study's implications extend development tailored strategies enhancing accuracy across different languages, contributing broader goal creating robust trustworthy systems global applications.

Язык: Английский

Процитировано

9

Enhancing Contextual Understanding of Mistral LLM with External Knowledge Bases DOI Creative Commons

Miyu Sasaki,

Natsumi Watanabe,

Tsukihito Komanaka

и другие.

Research Square (Research Square), Год журнала: 2024, Номер unknown

Опубликована: Апрель 5, 2024

Abstract This study explores the enhancement of contextual understanding and factual accuracy in Language Learning Models (LLMs), specifically Mistral LLM, through integration external knowledge bases. We developed a novel methodology for dynamically incorporating real-time information from diverse sources, aiming to address inherent limitations LLMs rooted their training datasets. Our experiments demonstrated significant improvements accuracy, precision, recall, F1 score, alongside qualitative enhancements response relevance accuracy. The research also tackled computational challenges integrating knowledge, ensuring model's efficiency practical applicability. work not only highlights potential bases augment capabilities but sets stage future advancements creating more intelligent, adaptable, contextually aware AI systems. findings contribute broader field NLP by offering insights into overcoming traditional LLMs, presenting step toward developing systems with enhanced real-world applicability accessibility.

Язык: Английский

Процитировано

8

Improving Generalization Beyond Training Data with Compositional Generalization in Large Language Models DOI Open Access

Wong Ho-tin,

Gar-lai Yip

Опубликована: Май 20, 2024

Enhancing compositional generalization in language models addresses a crucial challenge natural processing, significantly improving their ability to understand and generate novel combinations of known concepts. The investigation utilized the Mistral 7x8B model, employing advanced data augmentation refined training methodologies enhance performance. By incorporating diverse challenging compositions during training, model demonstrated substantial gains standard evaluation metrics, including accuracy, precision, recall, F1-score. Specialized metrics such as accuracy contextual coherence also showed marked improvement, reflecting model's enhanced capacity correct contextually relevant outputs when faced with compositions. study further highlighted significant reduction hallucination rates, underscoring increased logical consistency factual accuracy. This was statistically significant, indicating robust enhancement Qualitative analysis corroborated these findings, revealing more coherent narratives accurate information retrieval generated responses. These improvements are particularly important for real-world applications where reliability appropriateness essential. comprehensive effectiveness proposed techniques, providing valuable insights into underlying mechanisms that contribute improved findings underscore importance iterative experimentation validation refining architectures techniques. advancing capabilities models, this research contributes development robust, flexible, reliable AI systems capable handling broader range linguistic tasks greater understanding.

Язык: Английский

Процитировано

8

Designing Incremental Knowledge Enrichment in Generative Pre-trained Transformers DOI Creative Commons
Emilia A. Kowalczyk, Mateusz Nowakowski,

Z Brzezińska

и другие.

Research Square (Research Square), Год журнала: 2024, Номер unknown

Опубликована: Апрель 1, 2024

Abstract This article presents a novel approach to Incremental Knowledge Enrichment tailored for GPT-Neo, addressing the challenge of keeping Large Language Models (LLMs) updated with latest information without undergoing comprehensive retraining. We introduce dynamic linking mechanism that enables real-time integration diverse data sources, thereby enhancing model's accuracy, timeliness, and relevance. Through rigorous evaluation, our method demonstrates significant improvements in model performance across several metrics. The research contributes scalable efficient solution one most pressing issues AI, potentially revolutionizing maintenance applicability LLMs. findings underscore feasibility creating more adaptive, responsive, sustainable generative models, opening new avenues future advancements field.

Язык: Английский

Процитировано

7

Probabilistic Inference Layer Integration in Mistral LLM for Accurate Information Retrieval DOI Creative Commons
Bing Wang, Shiyu Wang, Qian Ouyang

и другие.

Research Square (Research Square), Год журнала: 2024, Номер unknown

Опубликована: Янв. 3, 2024

Abstract This study introduces a novel integration of Probabilistic Inference Layer (PIL) into the Mistral Large Language Model (LLM), aiming to address critical challenge accurate and reliable information retrieval in natural language processing. By employing advanced statistical models within PIL, enhanced LLM demonstrates marked improvement accuracy, context understanding, bias reduction. The PIL's application Bayesian networks sophisticated mathematical constructs, such as matrix calculus principles akin Bernoulli's Lorentz transformations, enables process with higher degree accuracy reliability. study's results indicate significant advancements model's performance across various tests, particularly discerning intent, reducing biases, handling complex logical operations. Despite its computational demands ongoing completely eliminate PIL establishes new benchmark for LLMs opens avenues future research. contributes field by demonstrating potential probabilistic methods enhancing capabilities generative AI models, thus paving way more AI-driven systems.

Язык: Английский

Процитировано

6

Reinforcement Learning for Optimized Information Retrieval in LLaMA DOI Creative Commons
Chien-Hung Tu, Hsien-Jung Hsu, Shih-Wen Chen

и другие.

Research Square (Research Square), Год журнала: 2024, Номер unknown

Опубликована: Янв. 10, 2024

Abstract This study introduces a new approach to enhance information retrieval accuracy in Large Language Models (LLMs) by integrating specially designed reinforcement learning algorithm into the LLaMA model. The research focuses on developing and implementing an that dynamically adapts model's response strategies user queries, based combination of dynamical systems theory relativistic physics. Empirical results demonstrate Optimized model exhibits significant improvements accuracy, relevance, coherence across various tasks compared Baseline LLaMA. advancement not only showcases potential realm natural language processing but also marks considerable step forward development AI capable nuanced understanding decision-making. study's findings have profound implications for future research, particularly enhancing practical applicability LLMs complex, real-world scenarios, set benchmark integration machine techniques models.

Язык: Английский

Процитировано

6

Augmenting Large Language Models with Reverse Proxy Style Retrieval Augmented Generation for Higher Factual Accuracy DOI Open Access

Po-hao Li,

Ya-yun Lai

Опубликована: Май 31, 2024

The increasing reliance on artificial intelligence for generating human-like text has brought attention to the critical issue of factual accuracy in language models. Introducing a novel approach, this research augments Llama model with reverse proxy-style Retrieval Augmented Generation (RAG) mechanism, significantly enhancing and coherence generated text. By dynamically incorporating relevant up-to-date information from diverse external data sources, RAG-augmented addresses inherent limitations static training data, thereby more reliable contextually appropriate responses. experimental evaluation, using precision, recall, F1-score, BLEU, ROUGE metrics, demonstrated substantial improvements, affirming effectiveness proposed system. findings reveal potential integrating retrieval mechanisms generative models achieve higher quality generation, offering valuable insights future practical applications fields where precision reliability are paramount.

Язык: Английский

Процитировано

6

Instruction Tuning on Large Language Models to Improve Reasoning Performance DOI Creative Commons
Emily Vaillancourt,

Christopher Thompson

Опубликована: Июнь 7, 2024

The growing demand for natural language processing models capable of understanding and executing complex instructions has driven significant advancements in model fine-tuning techniques. novel concept instruction tuning, which involves pre-trained on meticulously curated datasets, shown remarkable promise enhancing performance. research presented here focuses applying tuning to GPT2 (124M parameters) improve its reasoning capabilities the Multi-task Language Understanding (MMLU) dataset. By systematically curating a diverse set tasks corresponding instructions, rigorously model, improvements were achieved key performance metrics, including accuracy, precision, recall, F1-score. Experimental results demonstrated that instruction-tuned GPT-2 significantly outperformed baseline other stateof-the-art models, showcasing effectiveness approach. enhanced capacity follow detailed led more accurate contextually relevant responses, showing potential this methodology refine augment models. comprehensive preparation dataset iterative process critical factors achieving these substantial gains. study’s findings suggest can be powerful tool optimizing across variety domains, provided datasets are carefully validated. resulted model’s capabilities, as evidenced by metrics MMLU highlights an effective approach refining their applicability scenarios. demonstrating benefits prepared study provides valuable insights into technique further processing.

Язык: Английский

Процитировано

6