The Neural Response at the Fundamental Frequency of Speech is Modulated by Word-level Acoustic and Linguistic Information DOI Creative Commons
Mikolaj Kegler, Hugo Weissbart, Tobias Reichenbach

et al.

bioRxiv (Cold Spring Harbor Laboratory), Journal Year: 2022, Volume and Issue: unknown

Published: April 8, 2022

ABSTRACT Spoken language comprehension requires rapid and continuous integration of information, from lower-level acoustic to higher-level linguistic features. Much this processing occurs in the cerebral cortex. Its neural activity exhibits, for instance, correlates predictive processing, emerging at delays a few hundred milliseconds. However, auditory pathways are also characterized by extensive feedback loops cortical areas ones as well subcortical structures. Early can therefore be influenced cognitive processes, but it remains unclear whether such contributes processing. Here, we investigated early speech-evoked that emerges fundamental frequency. We analyzed EEG recordings obtained when subjects listened story read single speaker. identified response tracking speaker’s frequency occurred delay 11 ms, while another elicited high-frequency modulation envelope higher harmonics exhibited larger magnitude longer latency about 18 ms. Subsequently, determined these responses each individual word story. then quantified context-independent used model compute context-dependent surprisal precision. The represented how predictable is, given previous context, precision reflected confidence predicting next past context. found word-level were predominantly features: average its variability. Amongst features, only showed weak significant modulation. Our results show is already suggesting top-down response.

Language: Английский

Predictors for estimating subcortical EEG responses to continuous speech DOI Creative Commons
Joshua P. Kulasingham, Florine L. Bachmann, Kasper Eskelund

et al.

PLoS ONE, Journal Year: 2024, Volume and Issue: 19(2), P. e0297826 - e0297826

Published: Feb. 8, 2024

Perception of sounds and speech involves structures in the auditory brainstem that rapidly process ongoing stimuli. The role these processing can be investigated by measuring their electrical activity using scalp-mounted electrodes. However, typical analysis methods involve averaging neural responses to many short repetitive stimuli bear little relevance daily listening environments. Recently, subcortical more ecologically relevant continuous were detected linear encoding models. These estimate temporal response function (TRF), which is a regression model minimises error between measured signal predictor derived from stimulus. Using predictors highly non-linear peripheral system may improve TRF estimation accuracy peak detection. Here, we compare both simple complex models for estimating TRFs on electroencephalography (EEG) data 24 participants speech. We also investigate length required TRFs, find around 12 minutes sufficient clear wave V peaks (>3 dB SNR) seen nearly all participants. Interestingly, filterbank-based yield SNRs are not significantly different those estimated nerve, provided nonlinear effects adaptation appropriately modelled. Crucially, computing simpler than 50 times faster compared model. This work paves way efficient modelling detection speech, lead improved diagnosis metrics hearing impairment assistive technology.

Language: Английский

Citations

21

Subcortical responses to music and speech are alike while cortical responses diverge DOI Creative Commons
Shan Tong, Madeline S. Cappelloni, Ross K. Maddox

et al.

Scientific Reports, Journal Year: 2024, Volume and Issue: 14(1)

Published: Jan. 8, 2024

Music and speech are encountered daily unique to human beings. Both transformed by the auditory pathway from an initial acoustical encoding higher level cognition. Studies of cortex have revealed distinct brain responses music speech, but differences may emerge in or be inherited different subcortical encoding. In first part this study, we derived brainstem response (ABR), a measure encoding, recorded using two analysis methods. The method, described previously acoustically based, yielded very ABRs between sound classes. second however, developed here based on physiological model periphery, gave highly correlated speech. We determined superiority method through several metrics, suggesting there is no appreciable impact stimulus class (i.e., vs speech) way acoustics encoded subcortically. study's part, considered cortex. Our new resulted cortical becoming more similar with remaining differences. results taken together suggest that evidence for stimulus-class dependent processing at not level.

Language: Английский

Citations

12

Extending Subcortical EEG Responses to Continuous Speech to the Sound-Field DOI Creative Commons
Florine L. Bachmann, Joshua P. Kulasingham, Kasper Eskelund

et al.

Trends in Hearing, Journal Year: 2024, Volume and Issue: 28

Published: Jan. 1, 2024

The auditory brainstem response (ABR) is a valuable clinical tool for objective hearing assessment, which conventionally detected by averaging neural responses to thousands of short stimuli. Progressing beyond these unnatural stimuli, continuous speech presented via earphones have been recently using linear temporal functions (TRFs). Here, we extend earlier studies measuring subcortical in the sound-field, and assess amount data needed estimate TRFs. Electroencephalography (EEG) was recorded from 24 normal participants while they listened clicks stories loudspeakers. Subcortical TRFs were computed after accounting non-linear processing periphery either stimulus rectification or an nerve model. Our results demonstrated that could be reliably measured sound-field. estimated models outperformed simple rectification, 16 minutes sufficient all show clear wave V peaks both sound-field highly consistent earphone conditions, with click ABRs. However, required slightly more (16 minutes) achieve compared (12 minutes), possibly due effects room acoustics. By investigating this study lays groundwork bringing assessment closer real-life may lead improved evaluations smart technologies.

Language: Английский

Citations

4

Neural speech tracking in a virtual acoustic environment: audio-visual benefit for unscripted continuous speech DOI Creative Commons

Mareike Daeglau,

Jürgen Otten,

Giso Grimm

et al.

Frontiers in Human Neuroscience, Journal Year: 2025, Volume and Issue: 19

Published: April 9, 2025

The audio-visual benefit in speech perception-where congruent visual input enhances auditory processing-is well-documented across age groups, particularly challenging listening conditions and among individuals with varying hearing abilities. However, most studies rely on highly controlled laboratory environments scripted stimuli. Here, we examine the using unscripted, natural from untrained speakers within a virtual acoustic environment. Using electroencephalography (EEG) cortical tracking, assessed neural responses audio-visual, audio-only, visual-only, masked-lip to isolate role of lip movements. Additionally, analysed individual differences features speakers, including pitch, jitter, lip-openness, explore their influence tracking benefit. Results showed significant enhancement background noise, condition performing similarly audio-only condition, emphasizing importance movements adverse situations. Our findings reveal feasibility naturalistic stimuli underscore impact speaker characteristics integration real-world contexts.

Language: Английский

Citations

0

The neural response at the fundamental frequency of speech is modulated by word-level acoustic and linguistic information DOI Creative Commons
Mikolaj Kegler, Hugo Weissbart, Tobias Reichenbach

et al.

Frontiers in Neuroscience, Journal Year: 2022, Volume and Issue: 16

Published: July 22, 2022

Spoken language comprehension requires rapid and continuous integration of information, from lower-level acoustic to higher-level linguistic features. Much this processing occurs in the cerebral cortex. Its neural activity exhibits, for instance, correlates predictive processing, emerging at delays a few 100 ms. However, auditory pathways are also characterized by extensive feedback loops cortical areas ones as well subcortical structures. Early can therefore be influenced cognitive processes, but it remains unclear whether such contributes processing. Here, we investigated early speech-evoked that emerges fundamental frequency. We analyzed EEG recordings obtained when subjects listened story read single speaker. identified response tracking speaker's frequency occurred delay 11 ms, while another elicited high-frequency modulation envelope higher harmonics exhibited larger magnitude longer latency about 18 ms with an additional significant component around 40 Notably, earlier components likely originate structures, latter presumably involves contributions regions. Subsequently, determined these responses each individual word story. then quantified context-independent used model compute context-dependent surprisal precision. The represented how predictable is, given previous context, precision reflected confidence predicting next past context. found word-level were predominantly features: average its variability. Amongst features, only showed weak modulation. Our results show is already suggesting top-down response.

Language: Английский

Citations

18

Auditory encoding of natural speech at subcortical and cortical levels is not indicative of cognitive decline DOI Creative Commons
Elena Bolt, Nathalie Giroud

eNeuro, Journal Year: 2024, Volume and Issue: unknown, P. ENEURO.0545 - 23.2024

Published: April 24, 2024

More and more patients worldwide are diagnosed with dementia, which emphasizes the urgent need for early detection markers. In this study, we built on auditory hypersensitivity theory of a previous study—which postulated that responses to input in subcortex as well cortex enhanced cognitive decline—, examined encoding natural continuous speech at both neural levels its indicative potential decline. We recruited study participants aged 60 years older, who were divided into two groups based Montreal Cognitive Assessment, one group low scores (n = 19, signs decline), control 25). Participants completed an audiometric assessment then recorded their electroencephalography while they listened audiobook click sounds. derived Temporal Response Functions evoked potentials from data response amplitudes predict decline, controlling age hearing loss. Contrary our expectations, no evidence was observed decline; comparable groups. Moreover, combination showed predictive value These results challenge proposed hypothesis emphasize further research identify reliable markers Significance statement Research decline needs studies uncovering subtle changes pre-dementia stage. Neural processing have has not been widely explored studies. Here, first time, used natural, continuously spoken language examine older adults without quantified how brain tracks only cortical but also subcortical level. contrast suggesting found differences. believe represents important contribution search health old age.

Language: Английский

Citations

3

Neural encoding of linguistic speech cues is unaffected by cognitive decline, but decreases with increasing hearing impairment DOI Creative Commons
Elena Bolt, Nathalie Giroud

Scientific Reports, Journal Year: 2024, Volume and Issue: 14(1)

Published: Aug. 17, 2024

The multivariate temporal response function (mTRF) is an effective tool for investigating the neural encoding of acoustic and complex linguistic features in natural continuous speech. In this study, we investigated how representations speech derived from stimuli are related to early signs cognitive decline older adults, taking into account effects hearing. Participants without (

Language: Английский

Citations

2

Comparing Methods for Deriving the Auditory Brainstem Response to Continuous Speech in Human Listeners DOI Creative Commons
Shan Tong, Ross K. Maddox

bioRxiv (Cold Spring Harbor Laboratory), Journal Year: 2024, Volume and Issue: unknown

Published: May 30, 2024

Abstract Several tools have recently been developed to derive the Auditory Brainstem Response (ABR) from continuous natural speech, facilitating investigation into subcortical encoding of speech. These rely on deconvolution, which models auditory pathway as a linear system, where nonlinearly processed stimulus is taken input (i.e., regressor), electroencephalogram (EEG) data output, and ABR impulse response deconvolved recorded EEG regressor. In this study, we analyzed recordings subjects listening both unaltered speech synthesized “peaky speech.” We compared derived ABRs using three regressors: half-wave rectified (HWR) Maddox Lee (2018), glottal pulse train (GP) Polonenko (2021), nerve modeled (ANM) Shan et al. (2024). Our evaluation focused fidelity, efficiency, practicality each method in different scenarios. The results indicate that ANM regressor for peaky GP provided best performance, whereas HWR demonstrated relatively poorer performance. findings study will guide future research selecting most appropriate paradigm derivation continuous, naturalistic

Language: Английский

Citations

1

Level-Dependent Subcortical Electroencephalography Responses to Continuous Speech DOI Creative Commons
Joshua P. Kulasingham, Hamish Innes-Brown, Martin Enqvist

et al.

eNeuro, Journal Year: 2024, Volume and Issue: 11(8), P. ENEURO.0135 - 24.2024

Published: Aug. 1, 2024

The auditory brainstem response (ABR) is a measure of subcortical activity in to stimuli. wave V peak the ABR depends on stimulus intensity level, and has been widely used for clinical hearing assessment. Conventional methods estimate average electroencephalography (EEG) responses short unnatural stimuli such as clicks. Recent work moved toward more ecologically relevant continuous speech using linear deconvolution models called temporal functions (TRFs). Investigating whether TRF waveform changes with crucial step use natural assessments involving responses. Here, we develop level-dependent TRFs EEG data collected from 21 participants listening presented at 4 different levels. We find that can be detected almost all participants, are consistent click-ABR V. also investigate most suitable peripheral model generate predictors simple gammatone filterbanks perform best. Additionally, around 6 min may sufficient detecting effects peaks above noise floor segments higher intensity. Finally, show proof-of-concept even inherent fluctuations speech.

Language: Английский

Citations

1

The effect of auditory training on listening effort in hearing-aid users: insights from a pupillometry study DOI Creative Commons
Aleksandra Koprowska, Dorothea Wendt, Maja Šerman

et al.

International Journal of Audiology, Journal Year: 2024, Volume and Issue: unknown, P. 1 - 11

Published: Jan. 30, 2024

Objective The study investigated how auditory training affects effort exerted by hearing-impaired listeners in speech-in-noise task.

Language: Английский

Citations

1