Expert Systems with Applications, Год журнала: 2023, Номер 241, С. 122700 - 122700
Опубликована: Ноя. 29, 2023
Язык: Английский
Expert Systems with Applications, Год журнала: 2023, Номер 241, С. 122700 - 122700
Опубликована: Ноя. 29, 2023
Язык: Английский
Neurobiology of Language, Год журнала: 2024, Номер 5(1), С. 43 - 63
Опубликована: Янв. 1, 2024
Abstract Artificial neural networks have emerged as computationally plausible models of human language processing. A major criticism these is that the amount training data they receive far exceeds humans during learning. Here, we use two complementary approaches to ask how models’ ability capture fMRI responses sentences affected by data. First, evaluate GPT-2 trained on 1 million, 10 100 or billion words against an benchmark. We consider 100-million-word model be developmentally in terms given this similar what children are estimated exposed first years life. Second, test performance a 9-billion-token dataset reach state-of-the-art next-word prediction benchmark at different stages training. Across both approaches, find (i) already achieve near-maximal capturing sentences. Further, (ii) lower perplexity—a measure performance—is associated with stronger alignment data, suggesting received enough sufficiently high also acquire representations predictive responses. In tandem, findings establish although some necessary for ability, realistic (∼100 million words) may suffice.
Язык: Английский
Процитировано
17Annual Review of Neuroscience, Год журнала: 2024, Номер 47(1), С. 277 - 301
Опубликована: Апрель 26, 2024
It has long been argued that only humans could produce and understand language. But now, for the first time, artificial language models (LMs) achieve this feat. Here we survey new purchase LMs are providing on question of how is implemented in brain. We discuss why, a priori, might be expected to share similarities with human system. then summarize evidence represent linguistic information similarly enough enable relatively accurate brain encoding decoding during processing. Finally, examine which LM properties—their architecture, task performance, or training—are critical capturing neural responses review studies using as silico model organisms testing hypotheses about These ongoing investigations bring us closer understanding representations processes underlie our ability comprehend sentences express thoughts
Язык: Английский
Процитировано
13Current Opinion in Neurobiology, Год журнала: 2025, Номер 90, С. 102969 - 102969
Опубликована: Фев. 1, 2025
Язык: Английский
Процитировано
2NeuroImage, Год журнала: 2025, Номер unknown, С. 121253 - 121253
Опубликована: Май 1, 2025
Язык: Английский
Процитировано
0Asian Journal of Civil Engineering, Год журнала: 2023, Номер 25(3), С. 2367 - 2377
Опубликована: Ноя. 4, 2023
Язык: Английский
Процитировано
8Cognition, Год журнала: 2023, Номер 241, С. 105621 - 105621
Опубликована: Сен. 14, 2023
Deep neural networks (DNNs) are increasingly proposed as models of human vision, bolstered by their impressive performance on image classification and object recognition tasks. Yet, the extent to which DNNs capture fundamental aspects vision such color perception remains unclear. Here, we develop novel experiments for evaluating perceptual coherence embeddings in DNNs, assess how well these algorithms predict similarity judgments collected via an online survey. We find that state-of-the-art DNN architectures - including convolutional transformers provide strikingly diverge from (i) images with controlled properties, (ii) generated searches, (iii) real-world canonical CIFAR-10 dataset. compare against interpretable cognitively plausible model based wavelet decomposition, inspired foundational theories computational neuroscience. While one deep learning a trained style transfer task captures some perception, our algorithm provides more coherent better compared all examine. These results hold when altering high-level visual used train similar (e.g., versus segmentation), examining different layers given architecture. findings break new ground effort analyze representations machine improve ability serve vision. Implications learning, embodied cognition discussed.
Язык: Английский
Процитировано
7Neurobiology of Language, Год журнала: 2023, Номер 4(4), С. 611 - 636
Опубликована: Янв. 1, 2023
A fundamental question in neurolinguistics concerns the brain regions involved syntactic and semantic processing during speech comprehension, both at lexical (word processing) supra-lexical levels (sentence discourse processing). To what extent are these separated or intertwined? address this question, we introduce a novel approach exploiting neural language models to generate high-dimensional feature sets that separately encode information. More precisely, train model, GloVe, GPT-2, on text corpus from which selectively removed either We then assess features derived information-restricted still able predict fMRI time courses of humans listening naturalistic text. Furthermore, determine windows integration processing, manipulate size contextual information provided GPT-2. The analyses show that, while most comprehension sensitive features, relative magnitudes effects vary across regions. Moreover, best fitted by more spatially dissociated left hemisphere than right one, shows sensitivity longer contexts left. novelty our lies ability control for encoded models' embeddings manipulating training set. These "information-restricted" complement previous studies used probe bases language, shed new light its spatial organization.
Язык: Английский
Процитировано
6Опубликована: Июнь 11, 2024
Deep learning models trained on computer vision tasks are widely considered the most successful of human to date. The majority work that supports this idea evaluates how accurately these predict brain and behavioral responses static images objects natural scenes. Real-world vision, however, is highly dynamic, far less has focused evaluating accuracy deep in predicting stimuli move, involve more complicated, higher-order phenomena like social interactions. Here, we present a dataset videos captions involving complex multi-agent interactions, benchmark 350+ image, video, language neural videos. As with prior work, find many reach noise ceiling visual scene features along ventral stream (often primary substrate object recognition). In contrast, image poorly action interaction ratings lateral (a pathway increasingly theorized as specializing vision). Language (given sentence videos) better than either or video models, but they still perform at stream. Together results identify major gap AI's ability match highlight importance studying contexts.
Язык: Английский
Процитировано
2bioRxiv (Cold Spring Harbor Laboratory), Год журнала: 2024, Номер unknown
Опубликована: Дек. 26, 2024
Abstract Many artificial neural networks (ANNs) trained with ecologically plausible objectives on naturalistic data align behavior and representations in biological systems. Here, we show that this alignment is a consequence of convergence onto the same by high-performing ANNs brains. We developed method to identify stimuli systematically vary degree inter-model representation agreement. Across language vision, then showed from high-and low-agreement sets predictably modulated model-to-brain alignment. also examined which stimulus features distinguish high-from sentences images. Our results establish universality as core component provide new approach for using uncover structure computations.
Язык: Английский
Процитировано
2Proceedings of the National Academy of Sciences, Год журнала: 2024, Номер 121(25)
Опубликована: Июнь 14, 2024
This study unravels a concerning capability in Large Language Models (LLMs): the ability to understand and induce deception strategies. As LLMs like GPT-4 intertwine with human communication, aligning them values becomes paramount. ...Large language models (LLMs) are currently at forefront of intertwining AI systems communication everyday life. Thus, is great importance. However, given steady increase reasoning abilities, ...
Язык: Английский
Процитировано
1