Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 28.286
Filter
Add more filters

Publication year range
1.
Cell ; 184(18): 4626-4639.e13, 2021 09 02.
Article in English | MEDLINE | ID: mdl-34411517

ABSTRACT

Speech perception is thought to rely on a cortical feedforward serial transformation of acoustic into linguistic representations. Using intracranial recordings across the entire human auditory cortex, electrocortical stimulation, and surgical ablation, we show that cortical processing across areas is not consistent with a serial hierarchical organization. Instead, response latency and receptive field analyses demonstrate parallel and distinct information processing in the primary and nonprimary auditory cortices. This functional dissociation was also observed where stimulation of the primary auditory cortex evokes auditory hallucination but does not distort or interfere with speech perception. Opposite effects were observed during stimulation of nonprimary cortex in superior temporal gyrus. Ablation of the primary auditory cortex does not affect speech perception. These results establish a distributed functional organization of parallel information processing throughout the human auditory cortex and demonstrate an essential independent role for nonprimary auditory cortex in speech processing.


Subject(s)
Auditory Cortex/physiology , Speech/physiology , Audiometry, Pure-Tone , Electrodes , Electronic Data Processing , Humans , Phonetics , Pitch Perception , Reaction Time/physiology , Temporal Lobe/physiology
2.
Cell ; 174(1): 21-31.e9, 2018 06 28.
Article in English | MEDLINE | ID: mdl-29958109

ABSTRACT

In speech, the highly flexible modulation of vocal pitch creates intonation patterns that speakers use to convey linguistic meaning. This human ability is unique among primates. Here, we used high-density cortical recordings directly from the human brain to determine the encoding of vocal pitch during natural speech. We found neural populations in bilateral dorsal laryngeal motor cortex (dLMC) that selectively encoded produced pitch but not non-laryngeal articulatory movements. This neural population controlled short pitch accents to express prosodic emphasis on a word in a sentence. Other larynx cortical representations controlling voicing and longer pitch phrase contours were found at separate sites. dLMC sites also encoded vocal pitch during a non-speech singing task. Finally, direct focal stimulation of dLMC evoked laryngeal movements and involuntary vocalization, confirming its causal role in feedforward control. Together, these results reveal the neural basis for the voluntary control of vocal pitch in human speech. VIDEO ABSTRACT.


Subject(s)
Larynx/physiology , Motor Cortex/physiology , Speech , Adolescent , Adult , Brain Mapping , Electrocorticography , Female , Humans , Male , Middle Aged , Models, Biological , Young Adult
3.
Cell ; 164(6): 1269-1276, 2016 Mar 10.
Article in English | MEDLINE | ID: mdl-26967292

ABSTRACT

The use of vocalizations to communicate information and elaborate social bonds is an adaptation seen in many vertebrate species. Human speech is an extreme version of this pervasive form of communication. Unlike the vocalizations exhibited by the majority of land vertebrates, speech is a learned behavior requiring early sensory exposure and auditory feedback for its development and maintenance. Studies in humans and a small number of other species have provided insights into the neural and genetic basis for learned vocal communication and are helping to delineate the roles of brain circuits across the cortex, basal ganglia, and cerebellum in generating vocal behaviors. This Review provides an outline of the current knowledge about these circuits and the genes implicated in vocal communication, as well as a perspective on future research directions in this field.


Subject(s)
Speech , Vocalization, Animal , Animals , Brain/physiology , Forkhead Transcription Factors/genetics , Forkhead Transcription Factors/metabolism , Humans , Learning , Nervous System Diseases/genetics , Neural Pathways
4.
Nature ; 626(7999): 603-610, 2024 Feb.
Article in English | MEDLINE | ID: mdl-38297120

ABSTRACT

Humans are capable of generating extraordinarily diverse articulatory movement combinations to produce meaningful speech. This ability to orchestrate specific phonetic sequences, and their syllabification and inflection over subsecond timescales allows us to produce thousands of word sounds and is a core component of language1,2. The fundamental cellular units and constructs by which we plan and produce words during speech, however, remain largely unknown. Here, using acute ultrahigh-density Neuropixels recordings capable of sampling across the cortical column in humans, we discover neurons in the language-dominant prefrontal cortex that encoded detailed information about the phonetic arrangement and composition of planned words during the production of natural speech. These neurons represented the specific order and structure of articulatory events before utterance and reflected the segmentation of phonetic sequences into distinct syllables. They also accurately predicted the phonetic, syllabic and morphological components of upcoming words and showed a temporally ordered dynamic. Collectively, we show how these mixtures of cells are broadly organized along the cortical column and how their activity patterns transition from articulation planning to production. We also demonstrate how these cells reliably track the detailed composition of consonant and vowel sounds during perception and how they distinguish processes specifically related to speaking from those related to listening. Together, these findings reveal a remarkably structured organization and encoding cascade of phonetic representations by prefrontal neurons in humans and demonstrate a cellular process that can support the production of speech.


Subject(s)
Neurons , Phonetics , Prefrontal Cortex , Speech , Humans , Movement , Neurons/physiology , Speech/physiology , Speech Perception/physiology , Prefrontal Cortex/cytology , Prefrontal Cortex/physiology
5.
Nature ; 626(7999): 593-602, 2024 Feb.
Article in English | MEDLINE | ID: mdl-38093008

ABSTRACT

Understanding the neural basis of speech perception requires that we study the human brain both at the scale of the fundamental computational unit of neurons and in their organization across the depth of cortex. Here we used high-density Neuropixels arrays1-3 to record from 685 neurons across cortical layers at nine sites in a high-level auditory region that is critical for speech, the superior temporal gyrus4,5, while participants listened to spoken sentences. Single neurons encoded a wide range of speech sound cues, including features of consonants and vowels, relative vocal pitch, onsets, amplitude envelope and sequence statistics. Neurons at each cross-laminar recording exhibited dominant tuning to a primary speech feature while also containing a substantial proportion of neurons that encoded other features contributing to heterogeneous selectivity. Spatially, neurons at similar cortical depths tended to encode similar speech features. Activity across all cortical layers was predictive of high-frequency field potentials (electrocorticography), providing a neuronal origin for macroelectrode recordings from the cortical surface. Together, these results establish single-neuron tuning across the cortical laminae as an important dimension of speech encoding in human superior temporal gyrus.


Subject(s)
Auditory Cortex , Neurons , Speech Perception , Temporal Lobe , Humans , Acoustic Stimulation , Auditory Cortex/cytology , Auditory Cortex/physiology , Neurons/physiology , Phonetics , Speech , Speech Perception/physiology , Temporal Lobe/cytology , Temporal Lobe/physiology , Cues , Electrodes
6.
Nat Rev Neurosci ; 25(7): 473-492, 2024 Jul.
Article in English | MEDLINE | ID: mdl-38745103

ABSTRACT

Loss of speech after paralysis is devastating, but circumventing motor-pathway injury by directly decoding speech from intact cortical activity has the potential to restore natural communication and self-expression. Recent discoveries have defined how key features of speech production are facilitated by the coordinated activity of vocal-tract articulatory and motor-planning cortical representations. In this Review, we highlight such progress and how it has led to successful speech decoding, first in individuals implanted with intracranial electrodes for clinical epilepsy monitoring and subsequently in individuals with paralysis as part of early feasibility clinical trials to restore speech. We discuss high-spatiotemporal-resolution neural interfaces and the adaptation of state-of-the-art speech computational algorithms that have driven rapid and substantial progress in decoding neural activity into text, audible speech, and facial movements. Although restoring natural speech is a long-term goal, speech neuroprostheses already have performance levels that surpass communication rates offered by current assistive-communication technology. Given this accelerated rate of progress in the field, we propose key evaluation metrics for speed and accuracy, among others, to help standardize across studies. We finish by highlighting several directions to more fully explore the multidimensional feature space of speech and language, which will continue to accelerate progress towards a clinically viable speech neuroprosthesis.


Subject(s)
Brain-Computer Interfaces , Speech , Humans , Speech/physiology , Neural Prostheses , Animals
7.
Nature ; 620(7976): 1031-1036, 2023 Aug.
Article in English | MEDLINE | ID: mdl-37612500

ABSTRACT

Speech brain-computer interfaces (BCIs) have the potential to restore rapid communication to people with paralysis by decoding neural activity evoked by attempted speech into text1,2 or sound3,4. Early demonstrations, although promising, have not yet achieved accuracies sufficiently high for communication of unconstrained sentences from a large vocabulary1-7. Here we demonstrate a speech-to-text BCI that records spiking activity from intracortical microelectrode arrays. Enabled by these high-resolution recordings, our study participant-who can no longer speak intelligibly owing to amyotrophic lateral sclerosis-achieved a 9.1% word error rate on a 50-word vocabulary (2.7 times fewer errors than the previous state-of-the-art speech BCI2) and a 23.8% word error rate on a 125,000-word vocabulary (the first successful demonstration, to our knowledge, of large-vocabulary decoding). Our participant's attempted speech was decoded  at 62 words per minute, which is 3.4 times as fast as the previous record8 and begins to approach the speed of natural conversation (160 words per minute9). Finally, we highlight two aspects of the neural code for speech that are encouraging for speech BCIs: spatially intermixed tuning to speech articulators that makes accurate decoding possible from only a small region of cortex, and a detailed articulatory representation of phonemes that persists years after paralysis. These results show a feasible path forward for restoring rapid communication to people with paralysis who can no longer speak.


Subject(s)
Brain-Computer Interfaces , Neural Prostheses , Paralysis , Speech , Humans , Amyotrophic Lateral Sclerosis/physiopathology , Amyotrophic Lateral Sclerosis/rehabilitation , Cerebral Cortex/physiology , Microelectrodes , Paralysis/physiopathology , Paralysis/rehabilitation , Vocabulary
8.
Nature ; 620(7976): 1037-1046, 2023 Aug.
Article in English | MEDLINE | ID: mdl-37612505

ABSTRACT

Speech neuroprostheses have the potential to restore communication to people living with paralysis, but naturalistic speed and expressivity are elusive1. Here we use high-density surface recordings of the speech cortex in a clinical-trial participant with severe limb and vocal paralysis to achieve high-performance real-time decoding across three complementary speech-related output modalities: text, speech audio and facial-avatar animation. We trained and evaluated deep-learning models using neural data collected as the participant attempted to silently speak sentences. For text, we demonstrate accurate and rapid large-vocabulary decoding with a median rate of 78 words per minute and median word error rate of 25%. For speech audio, we demonstrate intelligible and rapid speech synthesis and personalization to the participant's pre-injury voice. For facial-avatar animation, we demonstrate the control of virtual orofacial movements for speech and non-speech communicative gestures. The decoders reached high performance with less than two weeks of training. Our findings introduce a multimodal speech-neuroprosthetic approach that has substantial promise to restore full, embodied communication to people living with severe paralysis.


Subject(s)
Face , Neural Prostheses , Paralysis , Speech , Humans , Cerebral Cortex/physiology , Cerebral Cortex/physiopathology , Clinical Trials as Topic , Communication , Deep Learning , Gestures , Movement , Neural Prostheses/standards , Paralysis/physiopathology , Paralysis/rehabilitation , Vocabulary , Voice
9.
Nat Rev Neurosci ; 24(11): 711-722, 2023 11.
Article in English | MEDLINE | ID: mdl-37783820

ABSTRACT

Is the singing voice processed distinctively in the human brain? In this Perspective, we discuss what might distinguish song processing from speech processing in light of recent work suggesting that some cortical neuronal populations respond selectively to song and we outline the implications for our understanding of auditory processing. We review the literature regarding the neural and physiological mechanisms of song production and perception and show that this provides evidence for key differences between song and speech processing. We conclude by discussing the significance of the notion that song processing is special in terms of how this might contribute to theories of the neurobiological origins of vocal communication and to our understanding of the neural circuitry underlying sound processing in the human cortex.


Subject(s)
Auditory Cortex , Humans , Auditory Perception/physiology , Speech/physiology , Brain/physiology , Acoustic Stimulation
10.
Nature ; 602(7895): 117-122, 2022 02.
Article in English | MEDLINE | ID: mdl-34987226

ABSTRACT

During conversation, people take turns speaking by rapidly responding to their partners while simultaneously avoiding interruption1,2. Such interactions display a remarkable degree of coordination, as gaps between turns are typically about 200 milliseconds3-approximately the duration of an eyeblink4. These latencies are considerably shorter than those observed in simple word-production tasks, which indicates that speakers often plan their responses while listening to their partners2. Although a distributed network of brain regions has been implicated in speech planning5-9, the neural dynamics underlying the specific preparatory processes that enable rapid turn-taking are poorly understood. Here we use intracranial electrocorticography to precisely measure neural activity as participants perform interactive tasks, and we observe a functionally and anatomically distinct class of planning-related cortical dynamics. We localize these responses to a frontotemporal circuit centred on the language-critical caudal inferior frontal cortex10 (Broca's region) and the caudal middle frontal gyrus-a region not normally implicated in speech planning11-13. Using a series of motor tasks, we then show that this planning network is more active when preparing speech as opposed to non-linguistic actions. Finally, we delineate planning-related circuitry during natural conversation that is nearly identical to the network mapped with our interactive tasks, and we find this circuit to be most active before participant speech during unconstrained turn-taking. Therefore, we have identified a speech planning network that is central to natural language generation during social interaction.


Subject(s)
Social Behavior , Speech/physiology , Adult , Aged , Broca Area/physiology , Electrocorticography , Executive Function , Female , Humans , Male , Middle Aged , Neural Pathways , Time Factors
11.
N Engl J Med ; 391(7): 609-618, 2024 Aug 15.
Article in English | MEDLINE | ID: mdl-39141853

ABSTRACT

BACKGROUND: Brain-computer interfaces can enable communication for people with paralysis by transforming cortical activity associated with attempted speech into text on a computer screen. Communication with brain-computer interfaces has been restricted by extensive training requirements and limited accuracy. METHODS: A 45-year-old man with amyotrophic lateral sclerosis (ALS) with tetraparesis and severe dysarthria underwent surgical implantation of four microelectrode arrays into his left ventral precentral gyrus 5 years after the onset of the illness; these arrays recorded neural activity from 256 intracortical electrodes. We report the results of decoding his cortical neural activity as he attempted to speak in both prompted and unstructured conversational contexts. Decoded words were displayed on a screen and then vocalized with the use of text-to-speech software designed to sound like his pre-ALS voice. RESULTS: On the first day of use (25 days after surgery), the neuroprosthesis achieved 99.6% accuracy with a 50-word vocabulary. Calibration of the neuroprosthesis required 30 minutes of cortical recordings while the participant attempted to speak, followed by subsequent processing. On the second day, after 1.4 additional hours of system training, the neuroprosthesis achieved 90.2% accuracy using a 125,000-word vocabulary. With further training data, the neuroprosthesis sustained 97.5% accuracy over a period of 8.4 months after surgical implantation, and the participant used it to communicate in self-paced conversations at a rate of approximately 32 words per minute for more than 248 cumulative hours. CONCLUSIONS: In a person with ALS and severe dysarthria, an intracortical speech neuroprosthesis reached a level of performance suitable to restore conversational communication after brief training. (Funded by the Office of the Assistant Secretary of Defense for Health Affairs and others; BrainGate2 ClinicalTrials.gov number, NCT00912041.).


Subject(s)
Amyotrophic Lateral Sclerosis , Brain-Computer Interfaces , Dysarthria , Speech , Humans , Male , Middle Aged , Amyotrophic Lateral Sclerosis/complications , Amyotrophic Lateral Sclerosis/rehabilitation , Calibration , Communication Aids for Disabled , Dysarthria/rehabilitation , Dysarthria/etiology , Electrodes, Implanted , Microelectrodes , Quadriplegia/etiology , Quadriplegia/rehabilitation
12.
PLoS Biol ; 22(2): e3002492, 2024 Feb.
Article in English | MEDLINE | ID: mdl-38386639

ABSTRACT

Stuttering occurs in early childhood during a dynamic phase of brain and behavioral development. The latest studies examining children at ages close to this critical developmental period have identified early brain alterations that are most likely linked to stuttering, while spontaneous recovery appears related to increased inter-area connectivity. By contrast, therapy-driven improvement in adults is associated with a functional reorganization within and beyond the speech network. The etiology of stuttering, however, remains enigmatic. This Unsolved Mystery highlights critical questions and points to neuroimaging findings that could inspire future research to uncover how genetics, interacting neural hierarchies, social context, and reward circuitry contribute to the many facets of stuttering.


Subject(s)
Stuttering , Child , Adult , Humans , Child, Preschool , Speech , Brain , Neuroimaging , Case-Control Studies
13.
PLoS Biol ; 22(10): e3002789, 2024 Oct.
Article in English | MEDLINE | ID: mdl-39352912

ABSTRACT

Within species, vocal and auditory systems presumably coevolved to converge on a critical temporal acoustic structure that can be best produced and perceived. While dogs cannot produce articulated sounds, they respond to speech, raising the question as to whether this heterospecific receptive ability could be shaped by exposure to speech or remains bounded by their own sensorimotor capacity. Using acoustic analyses of dog vocalisations, we show that their main production rhythm is slower than the dominant (syllabic) speech rate, and that human-dog-directed speech falls halfway in between. Comparative exploration of neural (electroencephalography) and behavioural responses to speech reveals that comprehension in dogs relies on a slower speech rhythm tracking (delta) than humans' (theta), even though dogs are equally sensitive to speech content and prosody. Thus, the dog audio-motor tuning differs from humans', and we hypothesise that humans may adjust their speech rate to this shared temporal channel as means to improve communication efficacy.


Subject(s)
Speech , Vocalization, Animal , Animals , Dogs , Humans , Vocalization, Animal/physiology , Speech/physiology , Male , Female , Electroencephalography , Auditory Perception/physiology , Adult , Human-Animal Interaction , Acoustic Stimulation , Speech Perception/physiology
14.
PLoS Biol ; 22(5): e3002631, 2024 May.
Article in English | MEDLINE | ID: mdl-38805517

ABSTRACT

Music and speech are complex and distinct auditory signals that are both foundational to the human experience. The mechanisms underpinning each domain are widely investigated. However, what perceptual mechanism transforms a sound into music or speech and how basic acoustic information is required to distinguish between them remain open questions. Here, we hypothesized that a sound's amplitude modulation (AM), an essential temporal acoustic feature driving the auditory system across processing levels, is critical for distinguishing music and speech. Specifically, in contrast to paradigms using naturalistic acoustic signals (that can be challenging to interpret), we used a noise-probing approach to untangle the auditory mechanism: If AM rate and regularity are critical for perceptually distinguishing music and speech, judging artificially noise-synthesized ambiguous audio signals should align with their AM parameters. Across 4 experiments (N = 335), signals with a higher peak AM frequency tend to be judged as speech, lower as music. Interestingly, this principle is consistently used by all listeners for speech judgments, but only by musically sophisticated listeners for music. In addition, signals with more regular AM are judged as music over speech, and this feature is more critical for music judgment, regardless of musical sophistication. The data suggest that the auditory system can rely on a low-level acoustic property as basic as AM to distinguish music from speech, a simple principle that provokes both neurophysiological and evolutionary experiments and speculations.


Subject(s)
Acoustic Stimulation , Auditory Perception , Music , Speech Perception , Humans , Male , Female , Adult , Auditory Perception/physiology , Acoustic Stimulation/methods , Speech Perception/physiology , Young Adult , Speech/physiology , Adolescent
15.
PLoS Biol ; 22(3): e3002534, 2024 Mar.
Article in English | MEDLINE | ID: mdl-38466713

ABSTRACT

Selective attention-related top-down modulation plays a significant role in separating relevant speech from irrelevant background speech when vocal attributes separating concurrent speakers are small and continuously evolving. Electrophysiological studies have shown that such top-down modulation enhances neural tracking of attended speech. Yet, the specific cortical regions involved remain unclear due to the limited spatial resolution of most electrophysiological techniques. To overcome such limitations, we collected both electroencephalography (EEG) (high temporal resolution) and functional magnetic resonance imaging (fMRI) (high spatial resolution), while human participants selectively attended to speakers in audiovisual scenes containing overlapping cocktail party speech. To utilise the advantages of the respective techniques, we analysed neural tracking of speech using the EEG data and performed representational dissimilarity-based EEG-fMRI fusion. We observed that attention enhanced neural tracking and modulated EEG correlates throughout the latencies studied. Further, attention-related enhancement of neural tracking fluctuated in predictable temporal profiles. We discuss how such temporal dynamics could arise from a combination of interactions between attention and prediction as well as plastic properties of the auditory cortex. EEG-fMRI fusion revealed attention-related iterative feedforward-feedback loops between hierarchically organised nodes of the ventral auditory object related processing stream. Our findings support models where attention facilitates dynamic neural changes in the auditory cortex, ultimately aiding discrimination of relevant sounds from irrelevant ones while conserving neural resources.


Subject(s)
Auditory Cortex , Speech Perception , Humans , Speech Perception/physiology , Speech , Feedback , Electroencephalography/methods , Auditory Cortex/physiology , Acoustic Stimulation/methods
16.
Proc Natl Acad Sci U S A ; 121(11): e2310766121, 2024 Mar 12.
Article in English | MEDLINE | ID: mdl-38442171

ABSTRACT

The neural correlates of sentence production are typically studied using task paradigms that differ considerably from the experience of speaking outside of an experimental setting. In this fMRI study, we aimed to gain a better understanding of syntactic processing in spontaneous production versus naturalistic comprehension in three regions of interest (BA44, BA45, and left posterior middle temporal gyrus). A group of participants (n = 16) was asked to speak about the events of an episode of a TV series in the scanner. Another group of participants (n = 36) listened to the spoken recall of a participant from the first group. To model syntactic processing, we extracted word-by-word metrics of phrase-structure building with a top-down and a bottom-up parser that make different hypotheses about the timing of structure building. While the top-down parser anticipates syntactic structure, sometimes before it is obvious to the listener, the bottom-up parser builds syntactic structure in an integratory way after all of the evidence has been presented. In comprehension, neural activity was found to be better modeled by the bottom-up parser, while in production, it was better modeled by the top-down parser. We additionally modeled structure building in production with two strategies that were developed here to make different predictions about the incrementality of structure building during speaking. We found evidence for highly incremental and anticipatory structure building in production, which was confirmed by a converging analysis of the pausing patterns in speech. Overall, this study shows the feasibility of studying the neural dynamics of spontaneous language production.


Subject(s)
Benchmarking , Mental Recall , Humans , Language , Software , Speech
17.
Proc Natl Acad Sci U S A ; 121(34): e2411167121, 2024 Aug 20.
Article in English | MEDLINE | ID: mdl-39136991

ABSTRACT

Evidence accumulates that the cerebellum's role in the brain is not restricted to motor functions. Rather, cerebellar activity seems to be crucial for a variety of tasks that rely on precise event timing and prediction. Due to its complex structure and importance in communication, human speech requires a particularly precise and predictive coordination of neural processes to be successfully comprehended. Recent studies proposed that the cerebellum is indeed a major contributor to speech processing, but how this contribution is achieved mechanistically remains poorly understood. The current study aimed to reveal a mechanism underlying cortico-cerebellar coordination and demonstrate its speech-specificity. In a reanalysis of magnetoencephalography data, we found that activity in the cerebellum aligned to rhythmic sequences of noise-vocoded speech, irrespective of its intelligibility. We then tested whether these "entrained" responses persist, and how they interact with other brain regions, when a rhythmic stimulus stopped and temporal predictions had to be updated. We found that only intelligible speech produced sustained rhythmic responses in the cerebellum. During this "entrainment echo," but not during rhythmic speech itself, cerebellar activity was coupled with that in the left inferior frontal gyrus, and specifically at rates corresponding to the preceding stimulus rhythm. This finding represents evidence for specific cerebellum-driven temporal predictions in speech processing and their relay to cortical regions.


Subject(s)
Cerebellum , Magnetoencephalography , Humans , Cerebellum/physiology , Male , Female , Adult , Speech Perception/physiology , Young Adult , Speech/physiology , Speech Intelligibility/physiology
18.
Proc Natl Acad Sci U S A ; 121(3): e2308837121, 2024 Jan 16.
Article in English | MEDLINE | ID: mdl-38198530

ABSTRACT

The development of individuality during learned behavior is a common trait observed across animal species; however, the underlying biological mechanisms remain understood. Similar to human speech, songbirds develop individually unique songs with species-specific traits through vocal learning. In this study, we investigate the developmental and molecular mechanisms underlying individuality in vocal learning by utilizing F1 hybrid songbirds (Taeniopygia guttata cross with Taeniopygia bichenovii), taking an integrating approach combining experimentally controlled systematic song tutoring, unbiased discriminant analysis of song features, and single-cell transcriptomics. When tutoring with songs from both parental species, F1 hybrid individuals exhibit evident diversity in their acquired songs. Approximately 30% of F1 hybrids selectively learn either song of the two parental species, while others develop merged songs that combine traits from both species. Vocal acoustic biases during vocal babbling initially appear as individual differences in songs among F1 juveniles and are maintained through the sensitive period of song vocal learning. These vocal acoustic biases emerge independently of the initial auditory experience of hearing the biological father's and passive tutored songs. We identify individual differences in transcriptional signatures in a subset of cell types, including the glutamatergic neurons projecting from the cortical vocal output nucleus to the hypoglossal nuclei, which are associated with variations of vocal acoustic features. These findings suggest that a genetically predisposed vocal motor bias serves as the initial origin of individual variation in vocal learning, influencing learning constraints and preferences.


Subject(s)
Individuality , Songbirds , Animals , Humans , Genetic Predisposition to Disease , Speech , Acoustics , Bias
19.
Proc Natl Acad Sci U S A ; 121(22): e2316149121, 2024 May 28.
Article in English | MEDLINE | ID: mdl-38768342

ABSTRACT

Speech impediments are a prominent yet understudied symptom of Parkinson's disease (PD). While the subthalamic nucleus (STN) is an established clinical target for treating motor symptoms, these interventions can lead to further worsening of speech. The interplay between dopaminergic medication, STN circuitry, and their downstream effects on speech in PD is not yet fully understood. Here, we investigate the effect of dopaminergic medication on STN circuitry and probe its association with speech and cognitive functions in PD patients. We found that changes in intrinsic functional connectivity of the STN were associated with alterations in speech functions in PD. Interestingly, this relationship was characterized by altered functional connectivity of the dorsolateral and ventromedial subdivisions of the STN with the language network. Crucially, medication-induced changes in functional connectivity between the STN's dorsolateral subdivision and key regions in the language network, including the left inferior frontal cortex and the left superior temporal gyrus, correlated with alterations on a standardized neuropsychological test requiring oral responses. This relation was not observed in the written version of the same test. Furthermore, changes in functional connectivity between STN and language regions predicted the medication's downstream effects on speech-related cognitive performance. These findings reveal a previously unidentified brain mechanism through which dopaminergic medication influences speech function in PD. Our study sheds light into the subcortical-cortical circuit mechanisms underlying impaired speech control in PD. The insights gained here could inform treatment strategies aimed at mitigating speech deficits in PD and enhancing the quality of life for affected individuals.


Subject(s)
Language , Parkinson Disease , Speech , Subthalamic Nucleus , Humans , Parkinson Disease/physiopathology , Parkinson Disease/drug therapy , Subthalamic Nucleus/physiopathology , Subthalamic Nucleus/drug effects , Male , Speech/physiology , Speech/drug effects , Female , Middle Aged , Aged , Magnetic Resonance Imaging , Dopamine/metabolism , Nerve Net/drug effects , Nerve Net/physiopathology , Cognition/drug effects , Dopamine Agents/pharmacology , Dopamine Agents/therapeutic use
20.
Annu Rev Neurosci ; 41: 527-552, 2018 07 08.
Article in English | MEDLINE | ID: mdl-29986161

ABSTRACT

How the cerebral cortex encodes auditory features of biologically important sounds, including speech and music, is one of the most important questions in auditory neuroscience. The pursuit to understand related neural coding mechanisms in the mammalian auditory cortex can be traced back several decades to the early exploration of the cerebral cortex. Significant progress in this field has been made in the past two decades with new technical and conceptual advances. This article reviews the progress and challenges in this area of research.


Subject(s)
Auditory Cortex/physiology , Auditory Pathways/physiology , Auditory Perception/physiology , Brain Mapping , Animals , Hearing , Humans , Music , Speech
SELECTION OF CITATIONS
SEARCH DETAIL