Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 92
Filtrar
Mais filtros

Base de dados
País/Região como assunto
Tipo de documento
Intervalo de ano de publicação
1.
PLoS Biol ; 20(7): e3001675, 2022 07.
Artigo em Inglês | MEDLINE | ID: mdl-35900975

RESUMO

The ability to recognize abstract features of voice during auditory perception is an intricate feat of human audition. For the listener, this occurs in near-automatic fashion to seamlessly extract complex cues from a highly variable auditory signal. Voice perception depends on specialized regions of auditory cortex, including superior temporal gyrus (STG) and superior temporal sulcus (STS). However, the nature of voice encoding at the cortical level remains poorly understood. We leverage intracerebral recordings across human auditory cortex during presentation of voice and nonvoice acoustic stimuli to examine voice encoding at the cortical level in 8 patient-participants undergoing epilepsy surgery evaluation. We show that voice selectivity increases along the auditory hierarchy from supratemporal plane (STP) to the STG and STS. Results show accurate decoding of vocalizations from human auditory cortical activity even in the complete absence of linguistic content. These findings show an early, less-selective temporal window of neural activity in the STG and STS followed by a sustained, strongly voice-selective window. Encoding models demonstrate divergence in the encoding of acoustic features along the auditory hierarchy, wherein STG/STS responses are best explained by voice category and acoustics, as opposed to acoustic features of voice stimuli alone. This is in contrast to neural activity recorded from STP, in which responses were accounted for by acoustic features. These findings support a model of voice perception that engages categorical encoding mechanisms within STG and STS to facilitate feature extraction.


Assuntos
Córtex Auditivo , Percepção da Fala , Voz , Estimulação Acústica , Córtex Auditivo/fisiologia , Percepção Auditiva/fisiologia , Mapeamento Encefálico/métodos , Humanos , Imageamento por Ressonância Magnética , Percepção da Fala/fisiologia , Lobo Temporal/fisiologia
2.
Psychol Sci ; 34(4): 468-480, 2023 04.
Artigo em Inglês | MEDLINE | ID: mdl-36791783

RESUMO

Categorization has a deep impact on behavior, but whether category learning is served by a single system or multiple systems remains debated. Here, we designed two well-equated nonspeech auditory category learning challenges to draw on putative procedural (information-integration) versus declarative (rule-based) learning systems among adult Hebrew-speaking control participants and individuals with dyslexia, a language disorder that has been linked to a selective disruption in the procedural memory system and in which phonological deficits are ubiquitous. We observed impaired information-integration category learning and spared rule-based category learning in the dyslexia group compared with the neurotypical group. Quantitative model-based analyses revealed reduced use of, and slower shifting to, optimal procedural-based strategies in dyslexia with hypothesis-testing strategy use on par with control participants. The dissociation is consistent with multiple category learning systems and points to the possibility that procedural learning inefficiencies across categories defined by complex, multidimensional exemplars may result in difficulty in phonetic category acquisition in dyslexia.


Assuntos
Dislexia , Aprendizagem , Adulto , Humanos , Fonética
3.
Ear Hear ; 43(1): 9-22, 2022.
Artigo em Inglês | MEDLINE | ID: mdl-34751676

RESUMO

Following a conversation in a crowded restaurant or at a lively party poses immense perceptual challenges for some individuals with normal hearing thresholds. A number of studies have investigated whether noise-induced cochlear synaptopathy (CS; damage to the synapses between cochlear hair cells and the auditory nerve following noise exposure that does not permanently elevate hearing thresholds) contributes to this difficulty. A few studies have observed correlations between proxies of noise-induced CS and speech perception in difficult listening conditions, but many have found no evidence of a relationship. To understand these mixed results, we reviewed previous studies that have examined noise-induced CS and performance on speech perception tasks in adverse listening conditions in adults with normal or near-normal hearing thresholds. Our review suggests that superficially similar speech perception paradigms used in previous investigations actually placed very different demands on sensory, perceptual, and cognitive processing. Speech perception tests that use low signal-to-noise ratios and maximize the importance of fine sensory details- specifically by using test stimuli for which lexical, syntactic, and semantic cues do not contribute to performance-are more likely to show a relationship to estimated CS levels. Thus, the current controversy as to whether or not noise-induced CS contributes to individual differences in speech perception under challenging listening conditions may be due in part to the fact that many of the speech perception tasks used in past studies are relatively insensitive to CS-induced deficits.


Assuntos
Percepção da Fala , Fala , Estimulação Acústica , Adulto , Limiar Auditivo/fisiologia , Humanos , Individualidade , Mascaramento Perceptivo , Percepção da Fala/fisiologia
4.
Proc Natl Acad Sci U S A ; 116(10): 4671-4680, 2019 03 05.
Artigo em Inglês | MEDLINE | ID: mdl-30782817

RESUMO

Humans are born as "universal listeners" without a bias toward any particular language. However, over the first year of life, infants' perception is shaped by learning native speech categories. Acoustically different sounds-such as the same word produced by different speakers-come to be treated as functionally equivalent. In natural environments, these categories often emerge incidentally without overt categorization or explicit feedback. However, the neural substrates of category learning have been investigated almost exclusively using overt categorization tasks with explicit feedback about categorization decisions. Here, we examined whether the striatum, previously implicated in category learning, contributes to incidental acquisition of sound categories. In the fMRI scanner, participants played a videogame in which sound category exemplars aligned with game actions and events, allowing sound categories to incidentally support successful game play. An experimental group heard nonspeech sound exemplars drawn from coherent category spaces, whereas a control group heard acoustically similar sounds drawn from a less structured space. Although the groups exhibited similar in-game performance, generalization of sound category learning and activation of the posterior striatum were significantly greater in the experimental than control group. Moreover, the experimental group showed brain-behavior relationships related to the generalization of all categories, while in the control group these relationships were restricted to the categories with structured sound distributions. Together, these results demonstrate that the striatum, through its interactions with the left superior temporal sulcus, contributes to incidental acquisition of sound category representations emerging from naturalistic learning environments.


Assuntos
Corpo Estriado/fisiologia , Aprendizagem , Estimulação Acústica , Percepção Auditiva , Corpo Estriado/diagnóstico por imagem , Feminino , Humanos , Imageamento por Ressonância Magnética , Masculino , Som , Jogos de Vídeo , Adulto Jovem
5.
J Acoust Soc Am ; 151(2): 992, 2022 02.
Artigo em Inglês | MEDLINE | ID: mdl-35232077

RESUMO

Speech contrasts are signaled by multiple acoustic dimensions, but these dimensions are not equally diagnostic. Moreover, the relative diagnosticity, or weight, of acoustic dimensions in speech can shift in different communicative contexts for both speech perception and speech production. However, the literature remains unclear on whether, and if so how, talkers adjust speech to emphasize different acoustic dimensions in the context of changing communicative demands. Here, we examine the interplay of flexible cue weights in speech production and perception across amplitude and duration, secondary non-spectral acoustic dimensions for phonated Mandarin Chinese lexical tone, across natural speech and whispering, which eliminates fundamental frequency contour, the primary acoustic dimension. Phonated and whispered Mandarin productions from native talkers revealed enhancement of both duration and amplitude cues in whispered, compared to phonated speech. When nonspeech amplitude-modulated noises modeled these patterns of enhancement, identification of the noises as Mandarin lexical tone categories was more accurate than identification of noises modeling phonated speech amplitude and duration cues. Thus, speakers exaggerate secondary cues in whispered speech and listeners make use of this information. Yet, enhancement is not symmetric among the four Mandarin lexical tones, indicating possible constraints on the realization of this enhancement.


Assuntos
Percepção da Fala , Fala , China , Sinais (Psicologia) , Fonética , Percepção da Altura Sonora
6.
Neuroimage ; 224: 117396, 2021 01 01.
Artigo em Inglês | MEDLINE | ID: mdl-32979522

RESUMO

To extract meaningful information from complex auditory scenes like a noisy playground, rock concert, or classroom, children can direct attention to different sound streams. One means of accomplishing this might be to align neural activity with the temporal structure of a target stream, such as a specific talker or melody. However, this may be more difficult for children with ADHD, who can struggle with accurately perceiving and producing temporal intervals. In this EEG study, we found that school-aged children's attention to one of two temporally-interleaved isochronous tone 'melodies' was linked to an increase in phase-locking at the melody's rate, and a shift in neural phase that aligned the neural responses with the attended tone stream. Children's attention task performance and neural phase alignment with the attended melody were linked to performance on temporal production tasks, suggesting that children with more robust control over motor timing were better able to direct attention to the time points associated with the target melody. Finally, we found that although children with ADHD performed less accurately on the tonal attention task than typically developing children, they showed the same degree of attentional modulation of phase locking and neural phase shifts, suggesting that children with ADHD may have difficulty with attentional engagement rather than attentional selection.


Assuntos
Transtorno do Deficit de Atenção com Hiperatividade/fisiopatologia , Córtex Auditivo/fisiopatologia , Percepção Auditiva/fisiologia , Som , Estimulação Acústica/métodos , Córtex Auditivo/fisiologia , Criança , Eletroencefalografia/métodos , Feminino , Humanos , Masculino
7.
J Int Neuropsychol Soc ; 27(1): 12-22, 2021 01.
Artigo em Inglês | MEDLINE | ID: mdl-32762781

RESUMO

OBJECTIVE: Acoustic distortions to the speech signal impair spoken language recognition, but healthy listeners exhibit adaptive plasticity consistent with rapid adjustments in how the distorted speech input maps to speech representations, perhaps through engagement of supervised error-driven learning. This puts adaptive plasticity in speech perception in an interesting position with regard to developmental dyslexia inasmuch as dyslexia impacts speech processing and may involve dysfunction in neurobiological systems hypothesized to be involved in adaptive plasticity. METHOD: Here, we examined typical young adult listeners (N = 17), and those with dyslexia (N = 16), as they reported the identity of native-language monosyllabic spoken words to which signal processing had been applied to create a systematic acoustic distortion. During training, all participants experienced incremental signal distortion increases to mildly distorted speech along with orthographic and auditory feedback indicating word identity following response across a brief, 250-trial training block. During pretest and posttest phases, no feedback was provided to participants. RESULTS: Word recognition across severely distorted speech was poor at pretest and equivalent across groups. Training led to improved word recognition for the most severely distorted speech at posttest, with evidence that adaptive plasticity generalized to support recognition of new tokens not previously experienced under distortion. However, training-related recognition gains for listeners with dyslexia were significantly less robust than for control listeners. CONCLUSIONS: Less efficient adaptive plasticity to speech distortions may impact the ability of individuals with dyslexia to deal with variability arising from sources like acoustic noise and foreign-accented speech.


Assuntos
Dislexia , Percepção da Fala , Adaptação Fisiológica , Percepção Auditiva , Humanos , Ruído , Adulto Jovem
8.
J Neurosci ; 39(14): 2698-2708, 2019 04 03.
Artigo em Inglês | MEDLINE | ID: mdl-30700532

RESUMO

The sensorimotor cortex is somatotopically organized to represent the vocal tract articulators such as lips, tongue, larynx, and jaw. How speech and articulatory features are encoded at the subcortical level, however, remains largely unknown. We analyzed LFP recordings from the subthalamic nucleus (STN) and simultaneous electrocorticography recordings from the sensorimotor cortex of 11 human subjects (1 female) with Parkinson's disease during implantation of deep-brain stimulation (DBS) electrodes while they read aloud three-phoneme words. The initial phonemes involved either articulation primarily with the tongue (coronal consonants) or the lips (labial consonants). We observed significant increases in high-gamma (60-150 Hz) power in both the STN and the sensorimotor cortex that began before speech onset and persisted for the duration of speech articulation. As expected from previous reports, in the sensorimotor cortex, the primary articulators involved in the production of the initial consonants were topographically represented by high-gamma activity. We found that STN high-gamma activity also demonstrated specificity for the primary articulator, although no clear topography was observed. In general, subthalamic high-gamma activity varied along the ventral-dorsal trajectory of the electrodes, with greater high-gamma power recorded in the dorsal locations of the STN. Interestingly, the majority of significant articulator-discriminative activity in the STN occurred before that in sensorimotor cortex. These results demonstrate that articulator-specific speech information is contained within high-gamma activity of the STN, but with different spatial and temporal organization compared with similar information encoded in the sensorimotor cortex.SIGNIFICANCE STATEMENT Clinical and electrophysiological evidence suggest that the subthalamic nucleus (STN) is involved in speech; however, this important basal ganglia node is ignored in current models of speech production. We previously showed that STN neurons differentially encode early and late aspects of speech production, but no previous studies have examined subthalamic functional organization for speech articulators. Using simultaneous LFP recordings from the sensorimotor cortex and the STN in patients with Parkinson's disease undergoing deep-brain stimulation surgery, we discovered that STN high-gamma activity tracks speech production at the level of vocal tract articulators before the onset of vocalization and often before related cortical encoding.


Assuntos
Mapeamento Encefálico/métodos , Eletrocorticografia/métodos , Estimulação Luminosa/métodos , Córtex Sensório-Motor/fisiologia , Fala/fisiologia , Núcleo Subtalâmico/fisiologia , Idoso , Feminino , Humanos , Masculino , Pessoa de Meia-Idade
9.
J Neurosci ; 38(24): 5620-5631, 2018 06 13.
Artigo em Inglês | MEDLINE | ID: mdl-29789378

RESUMO

Basal ganglia-thalamocortical loops mediate all motor behavior, yet little detail is known about the role of basal ganglia nuclei in speech production. Using intracranial recording during deep brain stimulation surgery in humans with Parkinson's disease, we tested the hypothesis that the firing rate of subthalamic nucleus neurons is modulated in sync with motor execution aspects of speech. Nearly half of 79 unit recordings exhibited firing-rate modulation during a syllable reading task across 12 subjects (male and female). Trial-to-trial timing of changes in subthalamic neuronal activity, relative to cue onset versus production onset, revealed that locking to cue presentation was associated more with units that decreased firing rate, whereas locking to speech onset was associated more with units that increased firing rate. These unique data indicate that subthalamic activity is dynamic during the production of speech, reflecting temporally-dependent inhibition and excitation of separate populations of subthalamic neurons.SIGNIFICANCE STATEMENT The basal ganglia are widely assumed to participate in speech production, yet no prior studies have reported detailed examination of speech-related activity in basal ganglia nuclei. Using microelectrode recordings from the subthalamic nucleus during a single-syllable reading task, in awake humans undergoing deep brain stimulation implantation surgery, we show that the firing rate of subthalamic nucleus neurons is modulated in response to motor execution aspects of speech. These results are the first to establish a role for subthalamic nucleus neurons in encoding of aspects of speech production, and they lay the groundwork for launching a modern subfield to explore basal ganglia function in human speech.


Assuntos
Neurônios/fisiologia , Fala/fisiologia , Núcleo Subtalâmico/fisiologia , Idoso , Idoso de 80 Anos ou mais , Feminino , Humanos , Masculino , Pessoa de Meia-Idade
10.
J Exp Child Psychol ; 188: 104673, 2019 12.
Artigo em Inglês | MEDLINE | ID: mdl-31430573

RESUMO

Adults outperform children on category learning that requires selective attention to individual dimensions (rule-based categories) due to their more highly developed working memory abilities, but much less is known about developmental differences in learning categories that require integration across multiple dimensions (information-integration categories). The current study investigated auditory information-integration category learning in 5- to 7-year-old children (n = 34) and 18- to 25-year-old adults (n = 35). Adults generally outperformed children during learning. However, some children learned the categories well and used strategies similar to those of adults, as assessed through decision-bound computational models. The results demonstrate that information-integration learning ability continues to develop throughout at least middle childhood. These results have implications for the development of mechanisms that contribute to speech category learning.


Assuntos
Formação de Conceito/fisiologia , Aprendizagem/fisiologia , Memória de Curto Prazo/fisiologia , Adolescente , Adulto , Atenção/fisiologia , Criança , Pré-Escolar , Feminino , Humanos , Inteligência/fisiologia , Masculino , Adulto Jovem
11.
J Acoust Soc Am ; 145(1): 131, 2019 01.
Artigo em Inglês | MEDLINE | ID: mdl-30710945

RESUMO

The automatic analysis of conversational audio remains difficult, in part, due to the presence of multiple talkers speaking in turns, often with significant intonation variations and overlapping speech. The majority of prior work on psychoacoustic speech analysis and system design has focused on single-talker speech or multi-talker speech with overlapping talkers (for example, the cocktail party effect). There has been much less focus on how listeners detect a change in talker or in probing the acoustic features significant in characterizing a talker's voice in conversational speech. This study examines human talker change detection (TCD) in multi-party speech utterances using a behavioral paradigm in which listeners indicate the moment of perceived talker change. Human reaction times in this task can be well-estimated by a model of the acoustic feature distance among speech segments before and after a change in talker, with estimation improving for models incorporating longer durations of speech prior to a talker change. Further, human performance is superior to several online and offline state-of-the-art machine TCD systems.


Assuntos
Processamento de Linguagem Natural , Percepção da Fala , Adulto , Feminino , Humanos , Masculino , Psicoacústica , Inteligibilidade da Fala , Interface para o Reconhecimento da Fala/normas
12.
J Neurosci ; 37(50): 12187-12201, 2017 12 13.
Artigo em Inglês | MEDLINE | ID: mdl-29109238

RESUMO

Auditory selective attention is vital in natural soundscapes. But it is unclear how attentional focus on the primary dimension of auditory representation-acoustic frequency-might modulate basic auditory functional topography during active listening. In contrast to visual selective attention, which is supported by motor-mediated optimization of input across saccades and pupil dilation, the primate auditory system has fewer means of differentially sampling the world. This makes spectrally-directed endogenous attention a particularly crucial aspect of auditory attention. Using a novel functional paradigm combined with quantitative MRI, we establish in male and female listeners that human frequency-band-selective attention drives activation in both myeloarchitectonically estimated auditory core, and across the majority of tonotopically mapped nonprimary auditory cortex. The attentionally driven best-frequency maps show strong concordance with sensory-driven maps in the same subjects across much of the temporal plane, with poor concordance in areas outside traditional auditory cortex. There is significantly greater activation across most of auditory cortex when best frequency is attended, versus ignored; the same regions do not show this enhancement when attending to the least-preferred frequency band. Finally, the results demonstrate that there is spatial correspondence between the degree of myelination and the strength of the tonotopic signal across a number of regions in auditory cortex. Strong frequency preferences across tonotopically mapped auditory cortex spatially correlate with R1-estimated myeloarchitecture, indicating shared functional and anatomical organization that may underlie intrinsic auditory regionalization.SIGNIFICANCE STATEMENT Perception is an active process, especially sensitive to attentional state. Listeners direct auditory attention to track a violin's melody within an ensemble performance, or to follow a voice in a crowded cafe. Although diverse pathologies reduce quality of life by impacting such spectrally directed auditory attention, its neurobiological bases are unclear. We demonstrate that human primary and nonprimary auditory cortical activation is modulated by spectrally directed attention in a manner that recapitulates its tonotopic sensory organization. Further, the graded activation profiles evoked by single-frequency bands are correlated with attentionally driven activation when these bands are presented in complex soundscapes. Finally, we observe a strong concordance in the degree of cortical myelination and the strength of tonotopic activation across several auditory cortical regions.


Assuntos
Atenção/fisiologia , Córtex Auditivo/fisiologia , Percepção Auditiva/fisiologia , Mapeamento Encefálico/métodos , Imageamento por Ressonância Magnética/métodos , Estimulação Acústica , Adulto , Córtex Auditivo/ultraestrutura , Sinais (Psicologia) , Feminino , Análise de Fourier , Humanos , Masculino , Pessoa de Meia-Idade , Bainha de Mielina/ultraestrutura , Reconhecimento Fisiológico de Modelo/fisiologia , Percepção da Altura Sonora/fisiologia , Espectrografia do Som
13.
Cereb Cortex ; 25(7): 1867-77, 2015 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-24451660

RESUMO

Human speech perception rapidly adapts to maintain comprehension under adverse listening conditions. For example, with exposure listeners can adapt to heavily accented speech produced by a non-native speaker. Outside the domain of speech perception, adaptive changes in sensory and motor processing have been attributed to cerebellar functions. The present functional magnetic resonance imaging study investigates whether adaptation in speech perception also involves the cerebellum. Acoustic stimuli were distorted using a vocoding plus spectral-shift manipulation and presented in a word recognition task. Regions in the cerebellum that showed differences before versus after adaptation were identified, and the relationship between activity during adaptation and subsequent behavioral improvements was examined. These analyses implicated the right Crus I region of the cerebellum in adaptive changes in speech perception. A functional correlation analysis with the right Crus I as a seed region probed for cerebral cortical regions with covarying hemodynamic responses during the adaptation period. The results provided evidence of a functional network between the cerebellum and language-related regions in the temporal and parietal lobes of the cerebral cortex. Consistent with known cerebellar contributions to sensorimotor adaptation, cerebro-cerebellar interactions may support supervised learning mechanisms that rely on sensory prediction error signals in speech perception.


Assuntos
Adaptação Fisiológica/fisiologia , Adaptação Psicológica/fisiologia , Cerebelo/fisiologia , Percepção da Fala/fisiologia , Estimulação Acústica/métodos , Mapeamento Encefálico , Circulação Cerebrovascular/fisiologia , Potenciais Evocados , Feminino , Humanos , Testes de Linguagem , Imageamento por Ressonância Magnética , Masculino , Testes Neuropsicológicos , Oxigênio/sangue , Reconhecimento Fisiológico de Modelo/fisiologia , Espectrografia do Som , Fala , Adulto Jovem
14.
Behav Brain Sci ; 37(2): 204-5, 2014 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-24775161

RESUMO

Speech is commonly claimed to relate to mirror neurons because of the alluring surface analogy of mirror neurons to the Motor Theory of speech perception, which posits that perception and production draw upon common motor-articulatory representations. We argue that the analogy fails and highlight examples of systems-level developmental approaches that have been more fruitful in revealing perception-production associations.


Assuntos
Evolução Biológica , Encéfalo/fisiologia , Aprendizagem/fisiologia , Neurônios-Espelho/fisiologia , Percepção Social , Animais , Humanos
15.
Curr Res Neurobiol ; 6: 100127, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38511174

RESUMO

The human voice is a critical stimulus for the auditory system that promotes social connection, informs the listener about identity and emotion, and acts as the carrier for spoken language. Research on voice processing in adults has informed our understanding of the unique status of the human voice in the mature auditory cortex and provided potential explanations for mechanisms that underly voice selectivity and identity processing. There is evidence that voice perception undergoes developmental change starting in infancy and extending through early adolescence. While even young infants recognize the voice of their mother, there is an apparent protracted course of development to reach adult-like selectivity for human voice over other sound categories and recognition of other talkers by voice. Gaps in the literature do not allow for an exact mapping of this trajectory or an adequate description of how voice processing and its neural underpinnings abilities evolve. This review provides a comprehensive account of developmental voice processing research published to date and discusses how this evidence fits with and contributes to current theoretical models proposed in the adult literature. We discuss how factors such as cognitive development, neural plasticity, perceptual narrowing, and language acquisition may contribute to the development of voice processing and its investigation in children. We also review evidence of voice processing abilities in premature birth, autism spectrum disorder, and phonagnosia to examine where and how deviations from the typical trajectory of development may manifest.

16.
Cogn Res Princ Implic ; 9(1): 35, 2024 Jun 05.
Artigo em Inglês | MEDLINE | ID: mdl-38834918

RESUMO

Multilingual speakers can find speech recognition in everyday environments like restaurants and open-plan offices particularly challenging. In a world where speaking multiple languages is increasingly common, effective clinical and educational interventions will require a better understanding of how factors like multilingual contexts and listeners' language proficiency interact with adverse listening environments. For example, word and phrase recognition is facilitated when competing voices speak different languages. Is this due to a "release from masking" from lower-level acoustic differences between languages and talkers, or higher-level cognitive and linguistic factors? To address this question, we created a "one-man bilingual cocktail party" selective attention task using English and Mandarin speech from one bilingual talker to reduce low-level acoustic cues. In Experiment 1, 58 listeners more accurately recognized English targets when distracting speech was Mandarin compared to English. Bilingual Mandarin-English listeners experienced significantly more interference and intrusions from the Mandarin distractor than did English listeners, exacerbated by challenging target-to-masker ratios. In Experiment 2, 29 Mandarin-English bilingual listeners exhibited linguistic release from masking in both languages. Bilinguals experienced greater release from masking when attending to English, confirming an influence of linguistic knowledge on the "cocktail party" paradigm that is separate from primarily energetic masking effects. Effects of higher-order language processing and expertise emerge only in the most demanding target-to-masker contexts. The "one-man bilingual cocktail party" establishes a useful tool for future investigations and characterization of communication challenges in the large and growing worldwide community of Mandarin-English bilinguals.


Assuntos
Atenção , Multilinguismo , Percepção da Fala , Humanos , Percepção da Fala/fisiologia , Adulto , Feminino , Masculino , Adulto Jovem , Atenção/fisiologia , Mascaramento Perceptivo/fisiologia , Psicolinguística
17.
bioRxiv ; 2024 Apr 05.
Artigo em Inglês | MEDLINE | ID: mdl-38617227

RESUMO

Prior lesion, noninvasive-imaging, and intracranial-electroencephalography (iEEG) studies have documented hierarchical, parallel, and distributed characteristics of human speech processing. Yet, there have not been direct, intracranial observations of the latency with which regions outside the temporal lobe respond to speech, or how these responses are impacted by task demands. We leveraged human intracranial recordings via stereo-EEG to measure responses from diverse forebrain sites during (i) passive listening to /bi/ and /pi/ syllables, and (ii) active listening requiring /bi/-versus-/pi/ categorization. We find that neural response latency increases from a few tens of ms in Heschl's gyrus (HG) to several tens of ms in superior temporal gyrus (STG), superior temporal sulcus (STS), and early parietal areas, and hundreds of ms in later parietal areas, insula, frontal cortex, hippocampus, and amygdala. These data also suggest parallel flow of speech information dorsally and ventrally, from HG to parietal areas and from HG to STG and STS, respectively. Latency data also reveal areas in parietal cortex, frontal cortex, hippocampus, and amygdala that are not responsive to the stimuli during passive listening but are responsive during categorization. Furthermore, multiple regions-spanning auditory, parietal, frontal, and insular cortices, and hippocampus and amygdala-show greater neural response amplitudes during active versus passive listening (a task-related effect). Overall, these results are consistent with hierarchical processing of speech at a macro level and parallel streams of information flow in temporal and parietal regions. These data also reveal regions where the speech code is stimulus-faithful and those that encode task-relevant representations.

18.
bioRxiv ; 2024 Mar 05.
Artigo em Inglês | MEDLINE | ID: mdl-37905141

RESUMO

Speech provides a rich context for exploring human cortical-basal ganglia circuit function, but direct intracranial recordings are rare. We recorded electrocorticographic signals in the cortex synchronously with single units in the subthalamic nucleus (STN), a basal ganglia node that receives direct input from widespread cortical regions, while participants performed a syllable repetition task during deep brain stimulation (DBS) surgery. We discovered that STN neurons exhibited spike-phase coupling (SPC) events with distinct combinations of frequency, location, and timing that indexed specific aspects of speech. The strength of SPC to posterior perisylvian cortex predicted phoneme production accuracy, while that of SPC to perirolandic cortex predicted time taken for articulation Thus, STN-cortical interactions are coordinated via transient bursts of behavior-specific synchronization that involves multiple neuronal populations and timescales. These results both suggest mechanisms that support auditory-sensorimotor integration during speech and explain why firing-rate based models are insufficient for explaining basal ganglia circuit behavior.

19.
bioRxiv ; 2024 May 26.
Artigo em Inglês | MEDLINE | ID: mdl-38826304

RESUMO

Efficient behavior is supported by humans' ability to rapidly recognize acoustically distinct sounds as members of a common category. Within auditory cortex, there are critical unanswered questions regarding the organization and dynamics of sound categorization. Here, we performed intracerebral recordings in the context of epilepsy surgery as 20 patient-participants listened to natural sounds. We built encoding models to predict neural responses using features of these sounds extracted from different layers within a sound-categorization deep neural network (DNN). This approach yielded highly accurate models of neural responses throughout auditory cortex. The complexity of a cortical site's representation (measured by the depth of the DNN layer that produced the best model) was closely related to its anatomical location, with shallow, middle, and deep layers of the DNN associated with core (primary auditory cortex), lateral belt, and parabelt regions, respectively. Smoothly varying gradients of representational complexity also existed within these regions, with complexity increasing along a posteromedial-to-anterolateral direction in core and lateral belt, and along posterior-to-anterior and dorsal-to-ventral dimensions in parabelt. When we estimated the time window over which each recording site integrates information, we found shorter integration windows in core relative to lateral belt and parabelt. Lastly, we found a relationship between the length of the integration window and the complexity of information processing within core (but not lateral belt or parabelt). These findings suggest hierarchies of timescales and processing complexity, and their interrelationship, represent a functional organizational principle of the auditory stream that underlies our perception of complex, abstract auditory information.

20.
J Exp Child Psychol ; 116(3): 728-37, 2013 Nov.
Artigo em Inglês | MEDLINE | ID: mdl-23827642

RESUMO

For both adults and children, acoustic context plays an important role in speech perception. For adults, both speech and nonspeech acoustic contexts influence perception of subsequent speech items, consistent with the argument that effects of context are due to domain-general auditory processes. However, prior research examining the effects of context on children's speech perception have focused on speech contexts; nonspeech contexts have not been explored previously. To better understand the developmental progression of children's use of contexts in speech perception and the mechanisms underlying that development, we created a novel experimental paradigm testing 5-year-old children's speech perception in several acoustic contexts. The results demonstrated that nonspeech context influences children's speech perception, consistent with claims that context effects arise from general auditory system properties rather than speech-specific mechanisms. This supports theoretical accounts of language development suggesting that domain-general processes play a role across the lifespan.


Assuntos
Acústica da Fala , Percepção da Fala , Estimulação Acústica , Adulto , Criança , Pré-Escolar , Feminino , Humanos , Masculino , Fonética , Inteligibilidade da Fala
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA