Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 96
Filtrar
1.
Front Neurosci ; 18: 1379988, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38784097

RESUMO

The prevalence of synthetic talking faces in both commercial and academic environments is increasing as the technology to generate them grows more powerful and available. While it has long been known that seeing the face of the talker improves human perception of speech-in-noise, recent studies have shown that synthetic talking faces generated by deep neural networks (DNNs) are also able to improve human perception of speech-in-noise. However, in previous studies the benefit provided by DNN synthetic faces was only about half that of real human talkers. We sought to determine whether synthetic talking faces generated by an alternative method would provide a greater perceptual benefit. The facial action coding system (FACS) is a comprehensive system for measuring visually discernible facial movements. Because the action units that comprise FACS are linked to specific muscle groups, synthetic talking faces generated by FACS might have greater verisimilitude than DNN synthetic faces which do not reference an explicit model of the facial musculature. We tested the ability of human observers to identity speech-in-noise accompanied by a blank screen; the real face of the talker; and synthetic talking faces generated either by DNN or FACS. We replicated previous findings of a large benefit for seeing the face of a real talker for speech-in-noise perception and a smaller benefit for DNN synthetic faces. FACS faces also improved perception, but only to the same degree as DNN faces. Analysis at the phoneme level showed that the performance of DNN and FACS faces was particularly poor for phonemes that involve interactions between the teeth and lips, such as /f/, /v/, and /th/. Inspection of single video frames revealed that the characteristic visual features for these phonemes were weak or absent in synthetic faces. Modeling the real vs. synthetic difference showed that increasing the realism of a few phonemes could substantially increase the overall perceptual benefit of synthetic faces.

2.
bioRxiv ; 2024 Feb 06.
Artigo em Inglês | MEDLINE | ID: mdl-38370798

RESUMO

The prevalence of synthetic talking faces in both commercial and academic environments is increasing as the technology to generate them grows more powerful and available. While it has long been known that seeing the face of the talker improves human perception of speech-in-noise, recent studies have shown that synthetic talking faces generated by deep neural networks (DNNs) are also able to improve human perception of speech-in-noise. However, in previous studies the benefit provided by DNN synthetic faces was only about half that of real human talkers. We sought to determine whether synthetic talking faces generated by an alternative method would provide a greater perceptual benefit. The facial action coding system (FACS) is a comprehensive system for measuring visually discernible facial movements. Because the action units that comprise FACS are linked to specific muscle groups, synthetic talking faces generated by FACS might have greater verisimilitude than DNN synthetic faces which do not reference an explicit model of the facial musculature. We tested the ability of human observers to identity speech-in-noise accompanied by a blank screen; the real face of the talker; and synthetic talking face generated either by DNN or FACS. We replicated previous findings of a large benefit for seeing the face of a real talker for speech-in-noise perception and a smaller benefit for DNN synthetic faces. FACS faces also improved perception, but only to the same degree as DNN faces. Analysis at the phoneme level showed that the performance of DNN and FACS faces was particularly poor for phonemes that involve interactions between the teeth and lips, such as /f/, /v/, and /th/. Inspection of single video frames revealed that the characteristic visual features for these phonemes were weak or absent in synthetic faces. Modeling the real vs. synthetic difference showed that increasing the realism of a few phonemes could substantially increase the overall perceptual benefit of synthetic faces, providing a roadmap for improving communication in this rapidly developing domain.

3.
Neuron ; 111(23): 3710-3715, 2023 Dec 06.
Artigo em Inglês | MEDLINE | ID: mdl-37944519

RESUMO

Sharing human brain data can yield scientific benefits, but because of various disincentives, only a fraction of these data is currently shared. We profile three successful data-sharing experiences from the NIH BRAIN Initiative Research Opportunities in Humans (ROH) Consortium and demonstrate benefits to data producers and to users.


Assuntos
Encéfalo , Neurofisiologia , Humanos , Disseminação de Informação
4.
eNeuro ; 10(10)2023 10.
Artigo em Inglês | MEDLINE | ID: mdl-37857509

RESUMO

Intracranial electroencephalography (iEEG) provides a unique opportunity to record and stimulate neuronal populations in the human brain. A key step in neuroscience inference from iEEG is localizing the electrodes relative to individual subject anatomy and identified regions in brain atlases. We describe a new software tool, Your Advanced Electrode Localizer (YAEL), that provides an integrated solution for every step of the electrode localization process. YAEL is compatible with all common data formats to provide an easy-to-use, drop-in replacement for problematic existing workflows that require users to grapple with multiple programs and interfaces. YAEL's automatic extrapolation and interpolation functions speed localization, especially important in patients with many implanted stereotactic (sEEG) electrode shafts. The graphical user interface is presented in a web browser for broad compatibility and includes an interactive 3D viewer for easier localization of nearby sEEG contacts. After localization is complete, users may enter or import data into YAEL's 3D viewer to create publication-ready visualizations of electrodes and brain anatomy, including identified brain areas from atlases; the response to experimental tasks measured with iEEG; and clinical measures such as epileptiform activity or the results of electrical stimulation mapping. YAEL is free and open source and does not depend on any commercial software. Installation instructions for Mac, Windows, and Linux are available at https://yael.wiki.


Assuntos
Eletrocorticografia , Eletroencefalografia , Humanos , Eletroencefalografia/métodos , Eletrocorticografia/métodos , Encéfalo/fisiologia , Mapeamento Encefálico/métodos , Eletrodos Implantados
5.
Sci Data ; 10(1): 719, 2023 10 19.
Artigo em Inglês | MEDLINE | ID: mdl-37857685

RESUMO

As data sharing has become more prevalent, three pillars - archives, standards, and analysis tools - have emerged as critical components in facilitating effective data sharing and collaboration. This paper compares four freely available intracranial neuroelectrophysiology data repositories: Data Archive for the BRAIN Initiative (DABI), Distributed Archives for Neurophysiology Data Integration (DANDI), OpenNeuro, and Brain-CODE. The aim of this review is to describe archives that provide researchers with tools to store, share, and reanalyze both human and non-human neurophysiology data based on criteria that are of interest to the neuroscientific community. The Brain Imaging Data Structure (BIDS) and Neurodata Without Borders (NWB) are utilized by these archives to make data more accessible to researchers by implementing a common standard. As the necessity for integrating large-scale analysis into data repository platforms continues to grow within the neuroscientific community, this article will highlight the various analytical and customizable tools developed within the chosen archives that may advance the field of neuroinformatics.


Assuntos
Disseminação de Informação , Neurofisiologia , Bases de Dados Factuais
6.
ArXiv ; 2023 Aug 30.
Artigo em Inglês | MEDLINE | ID: mdl-37426452

RESUMO

As data sharing has become more prevalent, three pillars - archives, standards, and analysis tools - have emerged as critical components in facilitating effective data sharing and collaboration. This paper compares four freely available intracranial neuroelectrophysiology data repositories: Data Archive for the BRAIN Initiative (DABI), Distributed Archives for Neurophysiology Data Integration (DANDI), OpenNeuro, and Brain-CODE. The aim of this review is to describe archives that provide researchers with tools to store, share, and reanalyze both human and non-human neurophysiology data based on criteria that are of interest to the neuroscientific community. The Brain Imaging Data Structure (BIDS) and Neurodata Without Borders (NWB) are utilized by these archives to make data more accessible to researchers by implementing a common standard. As the necessity for integrating large-scale analysis into data repository platforms continues to grow within the neuroscientific community, this article will highlight the various analytical and customizable tools developed within the chosen archives that may advance the field of neuroinformatics.

7.
Neuroimage ; 278: 120271, 2023 09.
Artigo em Inglês | MEDLINE | ID: mdl-37442310

RESUMO

Humans have the unique ability to decode the rapid stream of language elements that constitute speech, even when it is contaminated by noise. Two reliable observations about noisy speech perception are that seeing the face of the talker improves intelligibility and the existence of individual differences in the ability to perceive noisy speech. We introduce a multivariate BOLD fMRI measure that explains both observations. In two independent fMRI studies, clear and noisy speech was presented in visual, auditory and audiovisual formats to thirty-seven participants who rated intelligibility. An event-related design was used to sort noisy speech trials by their intelligibility. Individual-differences multidimensional scaling was applied to fMRI response patterns in superior temporal cortex and the dissimilarity between responses to clear speech and noisy (but intelligible) speech was measured. Neural dissimilarity was less for audiovisual speech than auditory-only speech, corresponding to the greater intelligibility of noisy audiovisual speech. Dissimilarity was less in participants with better noisy speech perception, corresponding to individual differences. These relationships held for both single word and entire sentence stimuli, suggesting that they were driven by intelligibility rather than the specific stimuli tested. A neural measure of perceptual intelligibility may aid in the development of strategies for helping those with impaired speech perception.


Assuntos
Percepção da Fala , Fala , Humanos , Imageamento por Ressonância Magnética , Individualidade , Percepção Visual/fisiologia , Percepção da Fala/fisiologia , Lobo Temporal/diagnóstico por imagem , Lobo Temporal/fisiologia , Inteligibilidade da Fala , Estimulação Acústica/métodos
8.
ACS Nano ; 17(10): 9442-9454, 2023 05 23.
Artigo em Inglês | MEDLINE | ID: mdl-37171407

RESUMO

MXenes are being heavily investigated in biomedical research, with applications ranging from regenerative medicine to bioelectronics. To enable the adoption and integration of MXenes into therapeutic platforms and devices, however, their stability under standard sterilization procedures must be established. Here, we present a comprehensive investigation of the electrical, chemical, structural, and mechanical effects of common thermal (autoclave) and chemical (ethylene oxide (EtO) and H2O2 gas plasma) sterilization protocols on both thin-film Ti3C2Tx MXene microelectrodes and mesoscale arrays made from Ti3C2Tx-infused cellulose-elastomer composites. We also evaluate the effectiveness of the sterilization processes in eliminating all pathogens from the Ti3C2Tx films and composites. Post-sterilization analysis revealed that autoclave and EtO did not alter the DC conductivity, electrochemical impedance, surface morphology, or crystallographic structure of Ti3C2Tx and were both effective at eliminating E. coli from both types of Ti3C2Tx-based devices. On the other end, exposure to H2O2 gas plasma sterilization for 45 min induced severe degradation of the structure and properties of Ti3C2Tx films and composites. The stability of the Ti3C2Tx after EtO and autoclave sterilization and the complete removal of pathogens establish the viability of both sterilization processes for Ti3C2Tx-based technologies.


Assuntos
Escherichia coli , Peróxido de Hidrogênio , Titânio/farmacologia , Esterilização
9.
Biometrics ; 79(2): 1226-1238, 2023 06.
Artigo em Inglês | MEDLINE | ID: mdl-35514244

RESUMO

This paper is motivated by studying differential brain activities to multiple experimental condition presentations in intracranial electroencephalography (iEEG) experiments. Contrasting effects of experimental conditions are often zero in most regions and nonzero in some local regions, yielding locally sparse functions. Such studies are essentially a function-on-scalar regression problem, with interest being focused not only on estimating nonparametric functions but also on recovering the function supports. We propose a weighted group bridge approach for simultaneous function estimation and support recovery in function-on-scalar mixed effect models, while accounting for heterogeneity present in functional data. We use B-splines to transform sparsity of functions to its sparse vector counterpart of increasing dimension, and propose a fast nonconvex optimization algorithm using nested alternative direction method of multipliers (ADMM) for estimation. Large sample properties are established. In particular, we show that the estimated coefficient functions are rate optimal in the minimax sense under the L2 norm and resemble a phase transition phenomenon. For support estimation, we derive a convergence rate under the L ∞ $L_{\infty }$ norm that leads to a selection consistency property under δ-sparsity, and obtain a result under strict sparsity using a simple sufficient regularity condition. An adjusted extended Bayesian information criterion is proposed for parameter tuning. The developed method is illustrated through simulations and an application to a novel iEEG data set to study multisensory integration.


Assuntos
Algoritmos , Encéfalo , Teorema de Bayes
10.
Brain Stimul ; 15(5): 1163-1177, 2022.
Artigo em Inglês | MEDLINE | ID: mdl-35985472

RESUMO

BACKGROUND: Direct electrical stimulation of early visual cortex evokes the perception of small spots of light known as phosphenes. Previous studies have examined the location, size, and brightness of phosphenes evoked by stimulation of single electrodes. While it has been envisioned that concurrent stimulation of many electrodes could be used as the basis for a visual cortical prosthesis, the percepts resulting from multi-electrode stimulation have not been fully characterized. OBJECTIVE: To understand the rules governing perception of phosphenes evoked by multi-electrode stimulation of visual cortex. METHODS: Multi-electrode stimulation was conducted in human epilepsy patients. We examined the number and spatial arrangement of phosphenes evoked by stimulation of individual multi-electrode groups (n = 8), and the ability of subjects to discriminate between the pattern of phosphenes generated by stimulation of different multi-electrode groups (n = 7). RESULTS: Simultaneous stimulation of pairs of electrodes separated by greater than 4 mm tended to produce perception of two distinct phosphenes. Simultaneous stimulation of three electrodes gave rise to a consistent spatial pattern of phosphenes, but with significant variation in the absolute location, size, and orientation of that pattern perceived on each trial. Although multi-electrode stimulation did not produce perception of recognizable forms, subjects could use the pattern of phosphenes evoked by stimulation to perform simple discriminations. CONCLUSIONS: The number of phosphenes produced by multi-electrode stimulation can be predicted using a model for spread of activity in early visual cortex, but there are additional subtle effects that must be accounted for.


Assuntos
Córtex Visual , Estimulação Elétrica , Eletrodos , Humanos , Fosfenos , Córtex Visual/fisiologia , Percepção Visual/fisiologia
11.
Neuron ; 110(2): 188-194, 2022 01 19.
Artigo em Inglês | MEDLINE | ID: mdl-35051364

RESUMO

Leveraging firsthand experience, BRAIN-funded investigators conducting intracranial human neuroscience research propose two fundamental ethical commitments: (1) maintaining the integrity of clinical care and (2) ensuring voluntariness. Principles, practices, and uncertainties related to these commitments are offered for future investigation.


Assuntos
Neurociências , Pesquisadores , Encéfalo , Humanos , Princípios Morais , Incerteza
12.
J Neurosci ; 42(6): 1054-1067, 2022 02 09.
Artigo em Inglês | MEDLINE | ID: mdl-34965979

RESUMO

Narrowband γ oscillations (NBG: ∼20-60 Hz) in visual cortex reflect rhythmic fluctuations in population activity generated by underlying circuits tuned for stimulus location, orientation, and color. A variety of theories posit a specific role for NBG in encoding and communicating this information within visual cortex. However, recent findings suggest a more nuanced role for NBG, given its dependence on certain stimulus feature configurations, such as coherent-oriented edges and specific hues. Motivated by these factors, we sought to quantify the independent and joint tuning properties of NBG to oriented and color stimuli using intracranial recordings from the human visual cortex (male and female). NBG was shown to display a cardinal orientation bias (horizontal) and also an end- and mid-spectral color bias (red/blue and green). When jointly probed, the cardinal bias for orientation was attenuated and an end-spectral preference for red and blue predominated. This loss of mid-spectral tuning occurred even for recording sites showing large responses to uniform green stimuli. Our results demonstrate the close, yet complex, link between the population dynamics driving NBG oscillations and known feature selectivity biases for orientation and color within visual cortex. Such a bias in stimulus tuning imposes new constraints on the functional significance of the visual γ rhythm. More generally, these biases in population electrophysiology will need to be considered in experiments using orientation or color features to examine the role of visual cortex in other domains, such as working memory and decision-making.SIGNIFICANCE STATEMENT Oscillations in electrophysiological activity occur in visual cortex in response to stimuli that strongly drive the orientation or color selectivity of visual neurons. The significance of this induced "γ rhythm" to brain function remains unclear. Answering this question requires understanding how and why some stimuli can reliably generate oscillatory γ activity while others do not. We examined how different orientations and colors independently and jointly modulate γ oscillations in the human brain. Our data show that γ oscillations are greatest for certain orientations and colors that reflect known response biases in visual cortex. Such findings complicate the functional significance of γ oscillations but open new avenues for linking circuits to population dynamics in visual cortex.


Assuntos
Percepção de Cores/fisiologia , Ritmo Gama/fisiologia , Orientação Espacial/fisiologia , Córtex Visual/fisiologia , Adulto , Eletrocorticografia , Feminino , Humanos , Masculino , Pessoa de Meia-Idade
13.
Neuroimage ; 247: 118796, 2022 02 15.
Artigo em Inglês | MEDLINE | ID: mdl-34906712

RESUMO

Regions of the human posterior superior temporal gyrus and sulcus (pSTG/S) respond to the visual mouth movements that constitute visual speech and the auditory vocalizations that constitute auditory speech, and neural responses in pSTG/S may underlie the perceptual benefit of visual speech for the comprehension of noisy auditory speech. We examined this possibility through the lens of multivoxel pattern responses in pSTG/S. BOLD fMRI data was collected from 22 participants presented with speech consisting of English sentences presented in five different formats: visual-only; auditory with and without added auditory noise; and audiovisual with and without auditory noise. Participants reported the intelligibility of each sentence with a button press and trials were sorted post-hoc into those that were more or less intelligible. Response patterns were measured in regions of the pSTG/S identified with an independent localizer. Noisy audiovisual sentences with very similar physical properties evoked very different response patterns depending on their intelligibility. When a noisy audiovisual sentence was reported as intelligible, the pattern was nearly identical to that elicited by clear audiovisual sentences. In contrast, an unintelligible noisy audiovisual sentence evoked a pattern like that of visual-only sentences. This effect was less pronounced for noisy auditory-only sentences, which evoked similar response patterns regardless of intelligibility. The successful integration of visual and auditory speech produces a characteristic neural signature in pSTG/S, highlighting the importance of this region in generating the perceptual benefit of visual speech.


Assuntos
Percepção Auditiva/fisiologia , Lobo Temporal/fisiologia , Percepção Visual/fisiologia , Estimulação Acústica , Adolescente , Adulto , Córtex Auditivo/fisiologia , Mapeamento Encefálico , Cognição , Compreensão/fisiologia , Feminino , Humanos , Processamento de Imagem Assistida por Computador , Imageamento por Ressonância Magnética , Masculino , Fala/fisiologia , Percepção da Fala/fisiologia , Adulto Jovem
14.
J Clin Invest ; 131(23)2021 12 01.
Artigo em Inglês | MEDLINE | ID: mdl-34850741

RESUMO

In this issue of the JCI, the dream of restoring useful vision to blind individuals with neurotechnology moves one step closer to realization. Fernández et al. implanted an electrode array with 96 penetrating electrodes in the visual cortex of a blind patient who had been without light perception for 16 years due to optic neuropathy. Remarkably, the patient was able to perceive visual patterns created by passing current through array electrodes. The use of a penetrating electrode array meant that action potentials from single neurons could be recorded to study the neural response to stimulation. Compared with electrodes resting on the cortical surface, penetrating electrodes require one-tenth the current to create a visual percept. However, patterned electrical stimulation often fails to produce the expected percept for penetrating and surface electrode arrays, highlighting the need for further research to untangle the relationship between stimulus and perception.


Assuntos
Córtex Visual , Próteses Visuais , Estimulação Elétrica , Eletrodos Implantados , Humanos , Nervo Óptico
15.
Curr Biol ; 31(9): R435-R437, 2021 05 10.
Artigo em Inglês | MEDLINE | ID: mdl-33974868

RESUMO

Primate brains contain specialized areas for perceiving social cues. New research shows that only some of these areas integrate visual faces with auditory voices.


Assuntos
Percepção Auditiva , Audição , Voz , Animais , Sinais (Psicologia) , Primatas
16.
Science ; 370(6521): 1168-1169, 2020 12 04.
Artigo em Inglês | MEDLINE | ID: mdl-33273090
17.
Cortex ; 133: 371-383, 2020 12.
Artigo em Inglês | MEDLINE | ID: mdl-33221701

RESUMO

The McGurk effect is a widely used measure of multisensory integration during speech perception. Two observations have raised questions about the validity of the effect as a tool for understanding speech perception. First, there is high variability in perception of the McGurk effect across different stimuli and observers. Second, across observers there is low correlation between McGurk susceptibility and recognition of visual speech paired with auditory speech-in-noise, another common measure of multisensory integration. Using the framework of the causal inference of multisensory speech (CIMS) model, we explored the relationship between the McGurk effect, syllable perception, and sentence perception in seven experiments with a total of 296 different participants. Perceptual reports revealed a relationship between the efficacy of different McGurk stimuli created from the same talker and perception of the auditory component of the McGurk stimuli presented in isolation, both with and without added noise. The CIMS model explained this strong stimulus-level correlation using the principles of noisy sensory encoding followed by optimal cue combination within a common representational space across speech types. Because the McGurk effect (but not speech-in-noise) requires the resolution of conflicting cues between modalities, there is an additional source of individual variability that can explain the weak observer-level correlation between McGurk and noisy speech. Power calculations show that detecting this weak correlation requires studies with many more participants than those conducted to-date. Perception of the McGurk effect and other types of speech can be explained by a common theoretical framework that includes causal inference, suggesting that the McGurk effect is a valid and useful experimental tool.


Assuntos
Ilusões , Percepção da Fala , Estimulação Acústica , Percepção Auditiva , Humanos , Estimulação Luminosa , Reconhecimento Psicológico , Fala , Percepção Visual
18.
J Neurosci ; 40(44): 8530-8542, 2020 10 28.
Artigo em Inglês | MEDLINE | ID: mdl-33023923

RESUMO

Natural conversation is multisensory: when we can see the speaker's face, visual speech cues improve our comprehension. The neuronal mechanisms underlying this phenomenon remain unclear. The two main alternatives are visually mediated phase modulation of neuronal oscillations (excitability fluctuations) in auditory neurons and visual input-evoked responses in auditory neurons. Investigating this question using naturalistic audiovisual speech with intracranial recordings in humans of both sexes, we find evidence for both mechanisms. Remarkably, auditory cortical neurons track the temporal dynamics of purely visual speech using the phase of their slow oscillations and phase-related modulations in broadband high-frequency activity. Consistent with known perceptual enhancement effects, the visual phase reset amplifies the cortical representation of concomitant auditory speech. In contrast to this, and in line with earlier reports, visual input reduces the amplitude of evoked responses to concomitant auditory input. We interpret the combination of improved phase tracking and reduced response amplitude as evidence for more efficient and reliable stimulus processing in the presence of congruent auditory and visual speech inputs.SIGNIFICANCE STATEMENT Watching the speaker can facilitate our understanding of what is being said. The mechanisms responsible for this influence of visual cues on the processing of speech remain incompletely understood. We studied these mechanisms by recording the electrical activity of the human brain through electrodes implanted surgically inside the brain. We found that visual inputs can operate by directly activating auditory cortical areas, and also indirectly by modulating the strength of cortical responses to auditory input. Our results help to understand the mechanisms by which the brain merges auditory and visual speech into a unitary perception.


Assuntos
Córtex Auditivo/fisiologia , Potenciais Evocados/fisiologia , Comunicação não Verbal/fisiologia , Adulto , Epilepsia Resistente a Medicamentos/cirurgia , Eletrocorticografia , Potenciais Evocados Auditivos/fisiologia , Potenciais Evocados Visuais/fisiologia , Feminino , Humanos , Pessoa de Meia-Idade , Neurônios/fisiologia , Comunicação não Verbal/psicologia , Estimulação Luminosa , Adulto Jovem
19.
Neuroimage ; 223: 117341, 2020 12.
Artigo em Inglês | MEDLINE | ID: mdl-32920161

RESUMO

Direct recording of neural activity from the human brain using implanted electrodes (iEEG, intracranial electroencephalography) is a fast-growing technique in human neuroscience. While the ability to record from the human brain with high spatial and temporal resolution has advanced our understanding, it generates staggering amounts of data: a single patient can be implanted with hundreds of electrodes, each sampled thousands of times a second for hours or days. The difficulty of exploring these vast datasets is the rate-limiting step in discovery. To overcome this obstacle, we created RAVE ("R Analysis and Visualization of iEEG"). All components of RAVE, including the underlying "R" language, are free and open source. User interactions occur through a web browser, making it transparent to the user whether the back-end data storage and computation are occurring locally, on a lab server, or in the cloud. Without writing a single line of computer code, users can create custom analyses, apply them to data from hundreds of iEEG electrodes, and instantly visualize the results on cortical surface models. Multiple types of plots are used to display analysis results, each of which can be downloaded as publication-ready graphics with a single click. RAVE consists of nearly 50,000 lines of code designed to prioritize an interactive user experience, reliability and reproducibility.


Assuntos
Encéfalo/fisiologia , Visualização de Dados , Eletroencefalografia , Processamento de Imagem Assistida por Computador/métodos , Eletrodos Implantados , Humanos , Reprodutibilidade dos Testes , Software
20.
J Neurosci ; 40(36): 6938-6948, 2020 09 02.
Artigo em Inglês | MEDLINE | ID: mdl-32727820

RESUMO

Experimentalists studying multisensory integration compare neural responses to multisensory stimuli with responses to the component modalities presented in isolation. This procedure is problematic for multisensory speech perception since audiovisual speech and auditory-only speech are easily intelligible but visual-only speech is not. To overcome this confound, we developed intracranial encephalography (iEEG) deconvolution. Individual stimuli always contained both auditory and visual speech, but jittering the onset asynchrony between modalities allowed for the time course of the unisensory responses and the interaction between them to be independently estimated. We applied this procedure to electrodes implanted in human epilepsy patients (both male and female) over the posterior superior temporal gyrus (pSTG), a brain area known to be important for speech perception. iEEG deconvolution revealed sustained positive responses to visual-only speech and larger, phasic responses to auditory-only speech. Confirming results from scalp EEG, responses to audiovisual speech were weaker than responses to auditory-only speech, demonstrating a subadditive multisensory neural computation. Leveraging the spatial resolution of iEEG, we extended these results to show that subadditivity is most pronounced in more posterior aspects of the pSTG. Across electrodes, subadditivity correlated with visual responsiveness, supporting a model in which visual speech enhances the efficiency of auditory speech processing in pSTG. The ability to separate neural processes may make iEEG deconvolution useful for studying a variety of complex cognitive and perceptual tasks.SIGNIFICANCE STATEMENT Understanding speech is one of the most important human abilities. Speech perception uses information from both the auditory and visual modalities. It has been difficult to study neural responses to visual speech because visual-only speech is difficult or impossible to comprehend, unlike auditory-only and audiovisual speech. We used intracranial encephalography deconvolution to overcome this obstacle. We found that visual speech evokes a positive response in the human posterior superior temporal gyrus, enhancing the efficiency of auditory speech processing.


Assuntos
Potenciais Evocados , Percepção da Fala , Lobo Temporal/fisiologia , Percepção Visual , Adulto , Eletrodos Implantados , Eletroencefalografia/instrumentação , Eletroencefalografia/métodos , Feminino , Humanos , Masculino
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA