Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 72
Filtrar
2.
Brain Commun ; 6(5): fcae346, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-39474046

RESUMEN

Convolutional neural networks (CNN) show great promise for translating decades of research on structural abnormalities in temporal lobe epilepsy into clinical practice. Three-dimensional CNNs typically outperform two-dimensional CNNs in medical imaging. Here we explore for the first time whether a three-dimensional CNN outperforms a two-dimensional CNN for identifying temporal lobe epilepsy-specific features on MRI. Using 1178 T1-weighted images (589 temporal lobe epilepsy, 589 healthy controls) from 12 surgical centres, we trained 3D and 2D CNNs for temporal lobe epilepsy versus healthy control classification, using feature visualization to identify important regions. The 3D CNN was compared to the 2D model and to a randomized model (comparison to chance). Further, we explored the effect of sample size with subsampling, examined model performance based on single-subject clinical characteristics, and tested the impact of image harmonization on model performance. Across 50 datapoints (10 runs with 5-folds each) the 3D CNN median accuracy was 86.4% (35.3% above chance) and the median F1-score was 86.1% (33.3% above chance). The 3D model yielded higher accuracy compared to the 2D model on 84% of datapoints (median 2D accuracy, 83.0%), a significant outperformance for the 3D model (binomial test: P < 0.001). This advantage of the 3D model was only apparent at the highest sample size. Saliency maps exhibited the importance of medial-ventral temporal, cerebellar, and midline subcortical regions across both models for classification. However, the 3D model had higher salience in the most important regions, the ventral-medial temporal and midline subcortical regions. Importantly, the model achieved high accuracy (82% accuracy) even in patients without MRI-identifiable hippocampal sclerosis. Finally, applying ComBat for harmonization did not improve performance. These findings highlight the value of 3D CNNs for identifying subtle structural abnormalities on MRI, especially in patients without clinically identified temporal lobe epilepsy lesions. Our findings also reveal that the advantage of 3D CNNs relies on large sample sizes for model training.

3.
Elife ; 132024 Sep 10.
Artículo en Inglés | MEDLINE | ID: mdl-39255194

RESUMEN

Across the animal kingdom, neural responses in the auditory cortex are suppressed during vocalization, and humans are no exception. A common hypothesis is that suppression increases sensitivity to auditory feedback, enabling the detection of vocalization errors. This hypothesis has been previously confirmed in non-human primates, however a direct link between auditory suppression and sensitivity in human speech monitoring remains elusive. To address this issue, we obtained intracranial electroencephalography (iEEG) recordings from 35 neurosurgical participants during speech production. We first characterized the detailed topography of auditory suppression, which varied across superior temporal gyrus (STG). Next, we performed a delayed auditory feedback (DAF) task to determine whether the suppressed sites were also sensitive to auditory feedback alterations. Indeed, overlapping sites showed enhanced responses to feedback, indicating sensitivity. Importantly, there was a strong correlation between the degree of auditory suppression and feedback sensitivity, suggesting suppression might be a key mechanism that underlies speech monitoring. Further, we found that when participants produced speech with simultaneous auditory feedback, posterior STG was selectively activated if participants were engaged in a DAF paradigm, suggesting that increased attentional load can modulate auditory feedback sensitivity.


The brain lowers its response to inputs we generate ourselves, such as moving or speaking. Essentially, our brain 'knows' what will happen next when we carry out these actions, and therefore does not need to react as strongly as it would to unexpected events. This is why we cannot tickle ourselves, and why the brain does not react as much to our own voice as it does to someone else's. Quieting down the brain's response also allows us to focus on things that are new or important without getting distracted by our own movements or sounds. Studies in non-human primates showed that neurons in the auditory cortex (the region of the brain responsible for processing sound) displayed suppressed levels of activity when the animals made sounds. Interestingly, when the primates heard an altered version of their own voice, many of these same neurons became more active. But it was unclear whether this also happens in humans. To investigate, Ozker et al. used a technique called electrocorticography to record neural activity in different regions of the human brain while participants spoke. The results showed that most areas of the brain involved in auditory processing showed suppressed activity when individuals were speaking. However, when people heard an altered version of their own voice which had an unexpected delay, those same areas displayed increased activity. In addition, Ozker et al. found that the higher the level of suppression in the auditory cortex, the more sensitive these areas were to changes in a person's speech. These findings suggest that suppressing the brain's response to self-generated speech may help in detecting errors during speech production. Speech deficits are common in various neurological disorders, such as stuttering, Parkinson's disease, and aphasia. Ozker et al. hypothesize that these deficits may arise because individuals fail to suppress activity in auditory regions of the brain, causing a struggle when detecting and correcting errors in their own speech. However, further experiments are needed to test this theory.


Asunto(s)
Retroalimentación Sensorial , Habla , Humanos , Masculino , Femenino , Adulto , Retroalimentación Sensorial/fisiología , Habla/fisiología , Adulto Joven , Corteza Auditiva/fisiología , Lóbulo Temporal/fisiología , Percepción del Habla/fisiología , Electroencefalografía , Electrocorticografía , Estimulación Acústica
4.
Epilepsia Open ; 9(5): 1931-1947, 2024 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-39225433

RESUMEN

OBJECTIVE: To characterize the experience of people with epilepsy and aligned healthcare workers (HCWs) during the first 18 months of the COVID-19 pandemic and compare experiences in high-income countries (HICs) with non-HICs. METHODS: Separate surveys for people with epilepsy and HCWs were distributed online in April 2020. Responses were collected to September 2021. Data were collected for COVID-19 infections, the effect of COVID-related restrictions, access to specialist help for epilepsy (people with epilepsy), and the impact of the pandemic on work productivity (HCWs). The frequency of responses for non-HICs and HICs were compared using non-parametric Chi-square tests. RESULTS: Two thousand one hundred and  five individuals with epilepsy from 53 countries and 392 HCWs from 26 countries provided data. The same proportion of people with epilepsy in non-HICs and HICs reported COVID-19 infection (7%). Those in HICs were more likely to report that COVID-19 measures had affected their health (32% vs. 23%; p < 0.001). There was no difference between non-HICs and HICs in the proportion who reported difficulty in obtaining help for epilepsy. HCWs in non-HICs were more likely to report COVID-19 infection than those in HICs (18% vs 6%; p = 0.001) and that their clinical work had been affected by concerns about contracting COVID-19, lack of personal protective equipment, and the impact of the pandemic on mental health (all p < 0.001). Compared to pre-pandemic practices, there was a significant shift to remote consultations in both non-HICs and HICs (p < 0.001). SIGNIFICANCE: While the frequency of COVID-19 infection was relatively low in these data from early in the pandemic, our findings suggest broader health consequences and an increased psychosocial burden, particularly among HCWs in non-HICs. Planning for future pandemics should prioritize mental healthcare alongside ensuring access to essential epilepsy services and expanding and enhancing access to remote consultations. PLAIN LANGUAGE SUMMARY: We asked people with epilepsy about the effects of COVID-19 on their health and healthcare. We wanted to compare responses from people in high-income countries and other countries. We found that people in high-income countries and other countries had similar levels of difficulty in getting help for their epilepsy. People in high-income countries were more likely to say that their general health had been affected. Healthcare workers in non-high-income settings were more likely to have contracted COVID-19 and have the care they deliver affected by the pandemic. Across all settings, COVID-19 associated with a large shift to remote consultations.


Asunto(s)
COVID-19 , Epilepsia , Personal de Salud , Humanos , COVID-19/epidemiología , Epilepsia/epidemiología , Femenino , Masculino , Adulto , Persona de Mediana Edad , Personal de Salud/psicología , Encuestas y Cuestionarios , Adulto Joven , Países Desarrollados , SARS-CoV-2 , Accesibilidad a los Servicios de Salud , Salud Global , Adolescente
5.
Neuron ; 112(18): 3211-3222.e5, 2024 Sep 25.
Artículo en Inglés | MEDLINE | ID: mdl-39096896

RESUMEN

Effective communication hinges on a mutual understanding of word meaning in different contexts. We recorded brain activity using electrocorticography during spontaneous, face-to-face conversations in five pairs of epilepsy patients. We developed a model-based coupling framework that aligns brain activity in both speaker and listener to a shared embedding space from a large language model (LLM). The context-sensitive LLM embeddings allow us to track the exchange of linguistic information, word by word, from one brain to another in natural conversations. Linguistic content emerges in the speaker's brain before word articulation and rapidly re-emerges in the listener's brain after word articulation. The contextual embeddings better capture word-by-word neural alignment between speaker and listener than syntactic and articulatory models. Our findings indicate that the contextual embeddings learned by LLMs can serve as an explicit numerical model of the shared, context-rich meaning space humans use to communicate their thoughts to one another.


Asunto(s)
Encéfalo , Electrocorticografía , Humanos , Encéfalo/fisiología , Masculino , Femenino , Lingüística , Epilepsia/fisiopatología , Adulto , Comunicación , Lenguaje , Modelos Neurológicos , Pensamiento/fisiología
6.
Epilepsy Res ; 206: 107425, 2024 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-39168079

RESUMEN

OBJECTIVE: We retrospectively explored patients with drug-resistant epilepsy (DRE) who previously underwent presurgical evaluation to identify correlations between surgical outcomes and pathogenic variants in epilepsy genes. METHODS: Through an international collaboration, we evaluated adult DRE patients who were screened for surgical candidacy. Patients with pathogenic (P) or likely pathogenic (LP) germline variants in genes relevant to their epilepsy were included, regardless of whether the genetic diagnosis was made before or after the presurgical evaluation. Patients were divided into two groups: resective surgery (RS) and non-resective surgery candidates (NRSC), with the latter group further divided into: palliative surgery (vagus nerve stimulation, deep brain stimulation, responsive neurostimulation or corpus callosotomy) and no surgery. We compared surgical candidacy evaluations and postsurgical outcomes in patients with different genetic abnormalities. RESULTS: We identified 142 patients with P/LP variants. After presurgical evaluation, 36 patients underwent RS, while 106 patients were NRSC. Patients with variants in ion channel and synaptic transmission genes were more common in the NRSC group (48 %), compared with the RS group (14 %) (p<0.001). Most patients in the RS group had tuberous sclerosis complex. Almost half (17/36, 47 %) in the RS group had Engel class I or II outcomes. Patients with channelopathies were less likely to undergo a surgical procedure than patients with mTORopathies, but when deemed suitable for resection had better surgical outcomes (71 % versus 41 % with Engel I/II). Within the NRSC group, 40 underwent palliative surgery, with 26/40 (65 %) having ≥50 % seizure reduction after mean follow-up of 11 years. Favourable palliative surgery outcomes were observed across a diverse range of genetic epilepsies. SIGNIFICANCE: Genomic findings, including a channelopathy diagnosis, should not preclude presurgical evaluation or epilepsy surgery, and appropriately selected cases may have good surgical outcomes. Prospective registries of patients with monogenic epilepsies who undergo epilepsy surgery can provide additional insights on outcomes.


Asunto(s)
Epilepsia Refractaria , Humanos , Epilepsia Refractaria/genética , Epilepsia Refractaria/cirugía , Femenino , Masculino , Adulto , Estudios Retrospectivos , Resultado del Tratamiento , Adulto Joven , Persona de Mediana Edad , Mutación de Línea Germinal/genética , Procedimientos Neuroquirúrgicos/métodos , Variación Genética/genética , Adolescente
7.
bioRxiv ; 2024 Oct 16.
Artículo en Inglés | MEDLINE | ID: mdl-39005394

RESUMEN

Recent research has used large language models (LLMs) to study the neural basis of naturalistic language processing in the human brain. LLMs have rapidly grown in complexity, leading to improved language processing capabilities. However, neuroscience researchers haven't kept up with the quick progress in LLM development. Here, we utilized several families of transformer-based LLMs to investigate the relationship between model size and their ability to capture linguistic information in the human brain. Crucially, a subset of LLMs were trained on a fixed training set, enabling us to dissociate model size from architecture and training set size. We used electrocorticography (ECoG) to measure neural activity in epilepsy patients while they listened to a 30-minute naturalistic audio story. We fit electrode-wise encoding models using contextual embeddings extracted from each hidden layer of the LLMs to predict word-level neural signals. In line with prior work, we found that larger LLMs better capture the structure of natural language and better predict neural activity. We also found a log-linear relationship where the encoding performance peaks in relatively earlier layers as model size increases. We also observed variations in the best-performing layer across different brain regions, corresponding to an organized language processing hierarchy.

8.
bioRxiv ; 2024 Jun 20.
Artículo en Inglés | MEDLINE | ID: mdl-38948730

RESUMEN

Syntax, the abstract structure of language, is a hallmark of human cognition. Despite its importance, its neural underpinnings remain obscured by inherent limitations of non-invasive brain measures and a near total focus on comprehension paradigms. Here, we address these limitations with high-resolution neurosurgical recordings (electrocorticography) and a controlled sentence production experiment. We uncover three syntactic networks that are broadly distributed across traditional language regions, but with focal concentrations in middle and inferior frontal gyri. In contrast to previous findings from comprehension studies, these networks process syntax mostly to the exclusion of words and meaning, supporting a cognitive architecture with a distinct syntactic system. Most strikingly, our data reveal an unexpected property of syntax: it is encoded independent of neural activity levels. We propose that this "low-activity coding" scheme represents a novel mechanism for encoding information, reserved for higher-order cognition more broadly.

9.
bioRxiv ; 2024 May 15.
Artículo en Inglés | MEDLINE | ID: mdl-38798614

RESUMEN

The ability to connect the form and meaning of a concept, known as word retrieval, is fundamental to human communication. While various input modalities could lead to identical word retrieval, the exact neural dynamics supporting this convergence relevant to daily auditory discourse remain poorly understood. Here, we leveraged neurosurgical electrocorticographic (ECoG) recordings from 48 patients and dissociated two key language networks that highly overlap in time and space integral to word retrieval. Using unsupervised temporal clustering techniques, we found a semantic processing network located in the middle and inferior frontal gyri. This network was distinct from an articulatory planning network in the inferior frontal and precentral gyri, which was agnostic to input modalities. Functionally, we confirmed that the semantic processing network encodes word surprisal during sentence perception. Our findings characterize how humans integrate ongoing auditory semantic information over time, a critical linguistic function from passive comprehension to daily discourse.

10.
PLoS Comput Biol ; 20(5): e1012161, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38815000

RESUMEN

Neural responses in visual cortex adapt to prolonged and repeated stimuli. While adaptation occurs across the visual cortex, it is unclear how adaptation patterns and computational mechanisms differ across the visual hierarchy. Here we characterize two signatures of short-term neural adaptation in time-varying intracranial electroencephalography (iEEG) data collected while participants viewed naturalistic image categories varying in duration and repetition interval. Ventral- and lateral-occipitotemporal cortex exhibit slower and prolonged adaptation to single stimuli and slower recovery from adaptation to repeated stimuli compared to V1-V3. For category-selective electrodes, recovery from adaptation is slower for preferred than non-preferred stimuli. To model neural adaptation we augment our delayed divisive normalization (DN) model by scaling the input strength as a function of stimulus category, enabling the model to accurately predict neural responses across multiple image categories. The model fits suggest that differences in adaptation patterns arise from slower normalization dynamics in higher visual areas interacting with differences in input strength resulting from category selectivity. Our results reveal systematic differences in temporal adaptation of neural population responses between lower and higher visual brain areas and show that a single computational model of history-dependent normalization dynamics, fit with area-specific parameters, accounts for these differences.


Asunto(s)
Adaptación Fisiológica , Modelos Neurológicos , Corteza Visual , Humanos , Corteza Visual/fisiología , Adaptación Fisiológica/fisiología , Adulto , Masculino , Femenino , Estimulación Luminosa , Biología Computacional , Adulto Joven , Electroencefalografía , Percepción Visual/fisiología , Electrocorticografía
11.
bioRxiv ; 2024 Sep 25.
Artículo en Inglés | MEDLINE | ID: mdl-38559163

RESUMEN

Objective: This study investigates speech decoding from neural signals captured by intracranial electrodes. Most prior works can only work with electrodes on a 2D grid (i.e., Electrocorticographic or ECoG array) and data from a single patient. We aim to design a deep-learning model architecture that can accommodate both surface (ECoG) and depth (stereotactic EEG or sEEG) electrodes. The architecture should allow training on data from multiple participants with large variability in electrode placements and the trained model should perform well on participants unseen during training. Approach: We propose a novel transformer-based model architecture named SwinTW that can work with arbitrarily positioned electrodes by leveraging their 3D locations on the cortex rather than their positions on a 2D grid. We train subject-specific models using data from a single participant and multi-patient models exploiting data from multiple participants. Main Results: The subject-specific models using only low-density 8x8 ECoG data achieved high decoding Pearson Correlation Coefficient with ground truth spectrogram (PCC=0.817), over N=43 participants, outperforming our prior convolutional ResNet model and the 3D Swin transformer model. Incorporating additional strip, depth, and grid electrodes available in each participant (N=39) led to further improvement (PCC=0.838). For participants with only sEEG electrodes (N=9), subject-specific models still enjoy comparable performance with an average PCC=0.798. The multi-subject models achieved high performance on unseen participants, with an average PCC=0.765 in leave-one-out cross-validation. Significance: The proposed SwinTW decoder enables future speech neuroprostheses to utilize any electrode placement that is clinically optimal or feasible for a particular participant, including using only depth electrodes, which are more routinely implanted in chronic neurosurgical procedures. Importantly, the generalizability of the multi-patient models suggests that such a model can be applied to new patients that do not have paired acoustic and neural data, providing an advance in neuroprostheses for people with speech disability, where acoustic-neural training data is not feasible.

12.
Brain Commun ; 6(2): fcae053, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38505231

RESUMEN

Cortical regions supporting speech production are commonly established using neuroimaging techniques in both research and clinical settings. However, for neurosurgical purposes, structural function is routinely mapped peri-operatively using direct electrocortical stimulation. While this method is the gold standard for identification of eloquent cortical regions to preserve in neurosurgical patients, there is lack of specificity of the actual underlying cognitive processes being interrupted. To address this, we propose mapping the temporal dynamics of speech arrest across peri-sylvian cortices by quantifying the latency between stimulation and speech deficits. In doing so, we are able to substantiate hypotheses about distinct region-specific functional roles (e.g. planning versus motor execution). In this retrospective observational study, we analysed 20 patients (12 female; age range 14-43) with refractory epilepsy who underwent continuous extra-operative intracranial EEG monitoring of an automatic speech task during clinical bedside language mapping. Latency to speech arrest was calculated as time from stimulation onset to speech arrest onset, controlling for individual speech rate. Most instances of motor-based arrest (87.5% of 96 instances) were in sensorimotor cortex with mid-range latencies to speech arrest with a distributional peak at 0.47 s. Speech arrest occurred in numerous regions, with relatively short latencies in supramarginal gyrus (0.46 s), superior temporal gyrus (0.51 s) and middle temporal gyrus (0.54 s), followed by relatively long latencies in sensorimotor cortex (0.72 s) and especially long latencies in inferior frontal gyrus (0.95 s). Non-parametric testing for speech arrest revealed that region predicted latency; latencies in supramarginal gyrus and in superior temporal gyrus were shorter than in sensorimotor cortex and in inferior frontal gyrus. Sensorimotor cortex is primarily responsible for motor-based arrest. Latencies to speech arrest in supramarginal gyrus and superior temporal gyrus (and to a lesser extent middle temporal gyrus) align with latencies to motor-based arrest in sensorimotor cortex. This pattern of relatively quick cessation of speech suggests that stimulating these regions interferes with the outgoing motor execution. In contrast, the latencies to speech arrest in inferior frontal gyrus and in ventral regions of sensorimotor cortex were significantly longer than those in temporoparietal regions. Longer latencies in the more frontal areas (including inferior frontal gyrus and ventral areas of precentral gyrus and postcentral gyrus) suggest that stimulating these areas interrupts a higher-level speech production process involved in planning. These results implicate the ventral specialization of sensorimotor cortex (including both precentral and postcentral gyri) for speech planning above and beyond motor execution.

13.
Nat Commun ; 15(1): 2768, 2024 Mar 30.
Artículo en Inglés | MEDLINE | ID: mdl-38553456

RESUMEN

Contextual embeddings, derived from deep language models (DLMs), provide a continuous vectorial representation of language. This embedding space differs fundamentally from the symbolic representations posited by traditional psycholinguistics. We hypothesize that language areas in the human brain, similar to DLMs, rely on a continuous embedding space to represent language. To test this hypothesis, we densely record the neural activity patterns in the inferior frontal gyrus (IFG) of three participants using dense intracranial arrays while they listened to a 30-minute podcast. From these fine-grained spatiotemporal neural recordings, we derive a continuous vectorial representation for each word (i.e., a brain embedding) in each patient. Using stringent zero-shot mapping we demonstrate that brain embeddings in the IFG and the DLM contextual embedding space have common geometric patterns. The common geometric patterns allow us to predict the brain embedding in IFG of a given left-out word based solely on its geometrical relationship to other non-overlapping words in the podcast. Furthermore, we show that contextual embeddings capture the geometry of IFG embeddings better than static word embeddings. The continuous brain embedding space exposes a vector-based neural code for natural language processing in the human brain.


Asunto(s)
Encéfalo , Lenguaje , Humanos , Corteza Prefrontal , Procesamiento de Lenguaje Natural
14.
Epilepsia ; 65(6): 1581-1588, 2024 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-38498313

RESUMEN

OBJECTIVE: New-onset refractory status epilepticus (NORSE) is a rare but severe clinical syndrome. Despite rigorous evaluation, the underlying cause is unknown in 30%-50% of patients and treatment strategies are largely empirical. The aim of this study was to describe clinical outcomes in a cohort of well-phenotyped, thoroughly investigated patients who survived the initial phase of cryptogenic NORSE managed in specialist centers. METHODS: Well-characterized cases of cryptogenic NORSE were identified through the EPIGEN and Critical Care EEG Monitoring Research Consortia (CCEMRC) during the period 2005-2019. Treating epileptologists reported on post-NORSE survival rates and sequelae in patients after discharge from hospital. Among survivors >6 months post-discharge, we report the rates and severity of active epilepsy, global disability, vocational, and global cognitive and mental health outcomes. We attempt to identify determinants of outcome. RESULTS: Among 48 patients who survived the acute phase of NORSE to the point of discharge from hospital, 9 had died at last follow-up, of whom 7 died within 6 months of discharge from the tertiary care center. The remaining 39 patients had high rates of active epilepsy as well as vocational, cognitive, and psychiatric comorbidities. The epilepsy was usually multifocal and typically drug resistant. Only a minority of patients had a good functional outcome. Therapeutic interventions were heterogenous during the acute phase of the illness. There was no clear relationship between the nature of treatment and clinical outcomes. SIGNIFICANCE: Among survivors of cryptogenic NORSE, longer-term outcomes in most patients were life altering and often catastrophic. Treatment remains empirical and variable. There is a pressing need to understand the etiology of cryptogenic NORSE and to develop tailored treatment strategies.


Asunto(s)
Epilepsia Refractaria , Estado Epiléptico , Sobrevivientes , Humanos , Masculino , Femenino , Adulto , Persona de Mediana Edad , Adulto Joven , Adolescente , Resultado del Tratamiento , Electroencefalografía , Niño
15.
bioRxiv ; 2024 Jun 21.
Artículo en Inglés | MEDLINE | ID: mdl-38370843

RESUMEN

Across the animal kingdom, neural responses in the auditory cortex are suppressed during vocalization, and humans are no exception. A common hypothesis is that suppression increases sensitivity to auditory feedback, enabling the detection of vocalization errors. This hypothesis has been previously confirmed in non-human primates, however a direct link between auditory suppression and sensitivity in human speech monitoring remains elusive. To address this issue, we obtained intracranial electroencephalography (iEEG) recordings from 35 neurosurgical participants during speech production. We first characterized the detailed topography of auditory suppression, which varied across superior temporal gyrus (STG). Next, we performed a delayed auditory feedback (DAF) task to determine whether the suppressed sites were also sensitive to auditory feedback alterations. Indeed, overlapping sites showed enhanced responses to feedback, indicating sensitivity. Importantly, there was a strong correlation between the degree of auditory suppression and feedback sensitivity, suggesting suppression might be a key mechanism that underlies speech monitoring. Further, we found that when participants produced speech with simultaneous auditory feedback, posterior STG was selectively activated if participants were engaged in a DAF paradigm, suggesting that increased attentional load can modulate auditory feedback sensitivity.

16.
bioRxiv ; 2024 Jan 17.
Artículo en Inglés | MEDLINE | ID: mdl-37745363

RESUMEN

Cortical regions supporting speech production are commonly established using neuroimaging techniques in both research and clinical settings. However, for neurosurgical purposes, structural function is routinely mapped peri-operatively using direct electrocortical stimulation. While this method is the gold standard for identification of eloquent cortical regions to preserve in neurosurgical patients, there is lack of specificity of the actual underlying cognitive processes being interrupted. To address this, we propose mapping the temporal dynamics of speech arrest across peri-sylvian cortices by quantifying the latency between stimulation and speech deficits. In doing so, we are able to substantiate hypotheses about distinct region-specific functional roles (e.g., planning versus motor execution). In this retrospective observational study, we analyzed 20 patients (12 female; age range 14-43) with refractory epilepsy who underwent continuous extra-operative intracranial EEG monitoring of an automatic speech task during clinical bedside language mapping. Latency to speech arrest was calculated as time from stimulation onset to speech arrest onset, controlling for individual speech rate. Most instances of motor-based arrest (87.5% of 96 instances) were in sensorimotor cortex with mid-range latencies to speech arrest with a distributional peak at 0.47 seconds. Speech arrest occurred in numerous regions, with relatively short latencies in supramarginal gyrus (0.46 seconds), superior temporal gyrus (0.51 seconds), and middle temporal gyrus (0.54 seconds), followed by relatively long latencies in sensorimotor cortex (0.72 seconds) and especially long latencies in inferior frontal gyrus (0.95 seconds). Nonparametric testing for speech arrest revealed that region predicted latency; latencies in supramarginal gyrus and in superior temporal gyrus were shorter than in sensorimotor cortex and in inferior frontal gyrus. Sensorimotor cortex is primarily responsible for motor-based arrest. Latencies to speech arrest in supramarginal gyrus and superior temporal gyrus (and to a lesser extent middle temporal gyrus) align with latencies to motor-based arrest in sensorimotor cortex. This pattern of relatively quick cessation of speech suggests that stimulating these regions interferes with the outgoing motor execution. In contrast, the latencies to speech arrest in inferior frontal gyrus and in ventral regions of sensorimotor cortex were significantly longer than those in temporoparietal regions. Longer latencies in the more frontal areas (including inferior frontal gyrus and ventral areas of precentral gyrus and postcentral gyrus) suggest that stimulating these areas interrupts a higher-level speech production process involved in planning. These results implicate the ventral specialization of sensorimotor cortex (including both precentral and postcentral gyri) for speech planning above and beyond motor execution.

17.
bioRxiv ; 2024 Feb 27.
Artículo en Inglés | MEDLINE | ID: mdl-37745548

RESUMEN

Neural responses in visual cortex adapt to prolonged and repeated stimuli. While adaptation occurs across the visual cortex, it is unclear how adaptation patterns and computational mechanisms differ across the visual hierarchy. Here we characterize two signatures of short-term neural adaptation in time-varying intracranial electroencephalography (iEEG) data collected while participants viewed naturalistic image categories varying in duration and repetition interval. Ventral- and lateral-occipitotemporal cortex exhibit slower and prolonged adaptation to single stimuli and slower recovery from adaptation to repeated stimuli compared to V1-V3. For category-selective electrodes, recovery from adaptation is slower for preferred than non-preferred stimuli. To model neural adaptation we augment our delayed divisive normalization (DN) model by scaling the input strength as a function of stimulus category, enabling the model to accurately predict neural responses across multiple image categories. The model fits suggest that differences in adaptation patterns arise from slower normalization dynamics in higher visual areas interacting with differences in input strength resulting from category selectivity. Our results reveal systematic differences in temporal adaptation of neural population responses across the human visual hierarchy and show that a single computational model of history-dependent normalization dynamics, fit with area-specific parameters, accounts for these differences.

18.
Epilepsia ; 65(2): 414-421, 2024 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-38060351

RESUMEN

OBJECTIVE: This study was undertaken to conduct external validation of previously published epilepsy surgery prediction tools using a large independent multicenter dataset and to assess whether these tools can stratify patients for being operated on and for becoming free of disabling seizures (International League Against Epilepsy stage 1 and 2). METHODS: We analyzed a dataset of 1562 patients, not used for tool development. We applied two scales: Epilepsy Surgery Grading Scale (ESGS) and Seizure Freedom Score (SFS); and two versions of Epilepsy Surgery Nomogram (ESN): the original version and the modified version, which included electroencephalographic data. For the ESNs, we used calibration curves and concordance indexes. We stratified the patients into three tiers for assessing the chances of attaining freedom from disabling seizures after surgery: high (ESGS = 1, SFS = 3-4, ESNs > 70%), moderate (ESGS = 2, SFS = 2, ESNs = 40%-70%), and low (ESGS = 2, SFS = 0-1, ESNs < 40%). We compared the three tiers as stratified by these tools, concerning the proportion of patients who were operated on, and for the proportion of patients who became free of disabling seizures. RESULTS: The concordance indexes for the various versions of the nomograms were between .56 and .69. Both scales (ESGS, SFS) and nomograms accurately stratified the patients for becoming free of disabling seizures, with significant differences among the three tiers (p < .05). In addition, ESGS and the modified ESN accurately stratified the patients for having been offered surgery, with significant difference among the three tiers (p < .05). SIGNIFICANCE: ESGS and the modified ESN (at thresholds of 40% and 70%) stratify patients undergoing presurgical evaluation into three tiers, with high, moderate, and low chance for favorable outcome, with significant differences between the groups concerning having surgery and becoming free of disabling seizures. Stratifying patients for epilepsy surgery has the potential to help select the optimal candidates in underprivileged areas and better allocate resources in developed countries.


Asunto(s)
Epilepsia , Humanos , Resultado del Tratamiento , Epilepsia/diagnóstico , Epilepsia/cirugía , Convulsiones/cirugía , Nomogramas , Medición de Riesgo
19.
Proc Natl Acad Sci U S A ; 120(42): e2300255120, 2023 10 17.
Artículo en Inglés | MEDLINE | ID: mdl-37819985

RESUMEN

Speech production is a complex human function requiring continuous feedforward commands together with reafferent feedback processing. These processes are carried out by distinct frontal and temporal cortical networks, but the degree and timing of their recruitment and dynamics remain poorly understood. We present a deep learning architecture that translates neural signals recorded directly from the cortex to an interpretable representational space that can reconstruct speech. We leverage learned decoding networks to disentangle feedforward vs. feedback processing. Unlike prevailing models, we find a mixed cortical architecture in which frontal and temporal networks each process both feedforward and feedback information in tandem. We elucidate the timing of feedforward and feedback-related processing by quantifying the derived receptive fields. Our approach provides evidence for a surprisingly mixed cortical architecture of speech circuitry together with decoding advances that have important implications for neural prosthetics.


Asunto(s)
Habla , Lóbulo Temporal , Humanos , Retroalimentación , Estimulación Acústica
20.
bioRxiv ; 2023 Sep 17.
Artículo en Inglés | MEDLINE | ID: mdl-37745380

RESUMEN

Decoding human speech from neural signals is essential for brain-computer interface (BCI) technologies restoring speech function in populations with neurological deficits. However, it remains a highly challenging task, compounded by the scarce availability of neural signals with corresponding speech, data complexity, and high dimensionality, and the limited publicly available source code. Here, we present a novel deep learning-based neural speech decoding framework that includes an ECoG Decoder that translates electrocorticographic (ECoG) signals from the cortex into interpretable speech parameters and a novel differentiable Speech Synthesizer that maps speech parameters to spectrograms. We develop a companion audio-to-audio auto-encoder consisting of a Speech Encoder and the same Speech Synthesizer to generate reference speech parameters to facilitate the ECoG Decoder training. This framework generates natural-sounding speech and is highly reproducible across a cohort of 48 participants. Among three neural network architectures for the ECoG Decoder, the 3D ResNet model has the best decoding performance (PCC=0.804) in predicting the original speech spectrogram, closely followed by the SWIN model (PCC=0.796). Our experimental results show that our models can decode speech with high correlation even when limited to only causal operations, which is necessary for adoption by real-time neural prostheses. We successfully decode speech in participants with either left or right hemisphere coverage, which could lead to speech prostheses in patients with speech deficits resulting from left hemisphere damage. Further, we use an occlusion analysis to identify cortical regions contributing to speech decoding across our models. Finally, we provide open-source code for our two-stage training pipeline along with associated preprocessing and visualization tools to enable reproducible research and drive research across the speech science and prostheses communities.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA