RESUMO
There is a rich tradition of research on the neuroanatomical correlates of spoken language production in aphasia using constrained tasks (e.g., picture naming), which offer controlled insights into the distinct processes that govern speech and language (i.e., lexical-semantic access, morphosyntactic construction, phonological encoding, speech motor programming/execution). Yet these tasks do not necessarily reflect everyday language use. In contrast, naturalistic language production (also referred to as connected speech or discourse) more closely approximates typical processing demands, requiring the dynamic integration of all aspects of speech and language. The brain bases of naturalistic language production remain relatively unknown, however, in part because of the difficulty in deriving features that are salient, quantifiable, and interpretable relative to both speech-language processes and the extant literature. The present cross-sectional observational study seeks to address these challenges by leveraging a validated and comprehensive auditory-perceptual measurement system that yields four explanatory dimensions of performance-Paraphasia (misselection of words and sounds), Logopenia (paucity of words), Agrammatism (grammatical omissions), and Motor speech (impaired speech motor programming/execution). We used this system to characterize naturalistic language production in a large and representative sample of individuals with acute post-stroke aphasia (n = 118). Scores on each of the four dimensions were correlated with lesion metrics, and multivariate associations among the dimensions and brain regions were then explored. Our findings revealed distinct yet overlapping neuroanatomical correlates throughout the left-hemisphere language network. Paraphasia and Logopenia were associated primarily with posterior regions, spanning both dorsal and ventral streams, which are critical for lexical-semantic access and phonological encoding. In contrast, Agrammatism and Motor speech were associated primarily with anterior regions of the dorsal stream that are involved in morphosyntactic construction and speech motor planning/execution respectively. Collectively, we view these results as constituting a brain-behavior model of naturalistic language production in aphasia, aligning with both historical and contemporary accounts of the neurobiology of spoken language production.
RESUMO
In the presence of temporally organized stimuli, there is a tendency to entrain to the beat, even at the neurological level. Previous research has shown that when adults listen to rhythmic stimuli and are asked to imagine the beat, their neural responses are the same as when the beat is physically accented. The current study explores the neural processing of simple beat structures where the beat is physically accented or inferred from a previously presented physically accented beat structure in a passive listening context. We further explore the associations of these neural correlates with behavioral and self-reported measures of musicality. Fifty-seven participants completed a passive listening EEG paradigm, a behavioral rhythm discrimination task, and a self-reported musicality questionnaire. Our findings suggest that when the beat is physically accented, individuals demonstrate distinct neural responses to the beat in the beta (13-23â Hz) and gamma (24-50â Hz) frequency bands. We further find that the neural marker in the beta band is associated with individuals' self-reported musical perceptual abilities. Overall, this study provides insights into the neural correlates of spontaneous beat processing and its connections with musicality.
Assuntos
Estimulação Acústica , Percepção Auditiva , Eletroencefalografia , Música , Humanos , Masculino , Feminino , Adulto Jovem , Percepção Auditiva/fisiologia , Adulto , Encéfalo/fisiologia , Potenciais Evocados Auditivos/fisiologia , Adolescente , Ritmo beta/fisiologiaRESUMO
Word deafness is a rare neurological disorder often observed following bilateral damage to superior temporal cortex and canonically defined as an auditory modality-specific deficit in word comprehension. The extent to which word deafness is dissociable from aphasia remains unclear given its heterogeneous presentation, and some have consequently posited that word deafness instead represents a stage in recovery from aphasia, where auditory and linguistic processing are affected to varying degrees and improve at differing rates. Here, we report a case of an individual (Mr. C) with bilateral temporal lobe lesions whose presentation evolved from a severe aphasia to an atypical form of word deafness, where auditory linguistic processing was impaired at the sentence level and beyond. We first reconstructed in detail Mr. C's stroke recovery through medical record review and supplemental interviewing. Then, using behavioral testing and multimodal neuroimaging, we documented a predominant auditory linguistic deficit in sentence and narrative comprehension-with markedly reduced behavioral performance and absent brain activation in the language network in the spoken modality exclusively. In contrast, Mr. C displayed near-unimpaired behavioral performance and robust brain activations in the language network for the linguistic processing of words, irrespective of modality. We argue that these findings not only support the view of word deafness as a stage in aphasia recovery but also further instantiate the important role of left superior temporal cortex in auditory linguistic processing.
Assuntos
Afasia , Surdez , Transtornos do Desenvolvimento da Linguagem , Acidente Vascular Cerebral , Humanos , Testes Neuropsicológicos , Afasia/etiologia , Acidente Vascular Cerebral/complicações , Lobo Temporal/patologia , Percepção AuditivaRESUMO
Introduction: The speech-to-song illusion is a robust effect where repeated speech induces the perception of singing; this effect has been extended to repeated excerpts of environmental sounds (sound-to-music effect). Here we asked whether repetition could elicit musical percepts in cochlear implant (CI) users, who experience challenges with perceiving music due to both physiological and device limitations. Methods: Thirty adult CI users and thirty age-matched controls with normal hearing (NH) completed two repetition experiments for speech and nonspeech sounds (water droplets). We hypothesized that CI users would experience the sound-to-music effect from temporal/rhythmic cues alone, but to a lesser magnitude compared to NH controls, given the limited access to spectral information CI users receive from their implants. Results: We found that CI users did experience the sound-to-music effect but to a lesser degree compared to NH participants. Musicality ratings were not associated with musical training or frequency resolution, and among CI users, clinical variables like duration of hearing loss also did not influence ratings. Discussion: Cochlear implants provide a strong clinical model for disentangling the effects of spectral and temporal information in an acoustic signal; our results suggest that temporal cues are sufficient to perceive the sound-to-music effect when spectral resolution is limited. Additionally, incorporating short repetitions into music specially designed for CI users may provide a promising way for them to experience music.
RESUMO
Purpose: Community aphasia groups serve an important purpose in enhancing the quality of life and psychosocial well-being of individuals with chronic aphasia. Here, we describe the Aphasia Group of Middle Tennessee, a community aphasia group with a 17-year (and continuing) history, housed within Vanderbilt University Medical Center in Nashville, Tennessee. Method: We describe in detail the history, philosophy, design, curriculum, and facilitation model of this group. We also present both quantitative and qualitative outcomes from group members and their loved ones. Results: Group members and their loved ones alike indicated highly positive assessments of the format and value of the Aphasia Group of Middle Tennessee. Conclusion: By characterizing in detail the successful Aphasia Group of Middle Tennessee, we hope this can serve as a model for clinicians interested in starting their own community aphasia groups, in addition to reaching individuals living with chronic aphasia and their loved ones through the accessible and aphasia-friendly materials provided with this clinical focus article.
RESUMO
We conducted a systematic review and meta-analysis of 30 functional magnetic resonance imaging studies investigating processing of musical rhythms in neurotypical adults. First, we identified a general network for musical rhythm, encompassing all relevant sensory and motor processes (Beat-based, rest baseline, 12 contrasts) which revealed a large network involving auditory and motor regions. This network included the bilateral superior temporal cortices, supplementary motor area (SMA), putamen, and cerebellum. Second, we identified more precise loci for beat-based musical rhythms (Beat-based, audio-motor control, 8 contrasts) in the bilateral putamen. Third, we identified regions modulated by beat based rhythmic complexity (Complexity, 16 contrasts) which included the bilateral SMA-proper/pre-SMA, cerebellum, inferior parietal regions, and right temporal areas. This meta-analysis suggests that musical rhythm is largely represented in a bilateral cortico-subcortical network. Our findings align with existing theoretical frameworks about auditory-motor coupling to a musical beat and provide a foundation for studying how the neural bases of musical rhythm may overlap with other cognitive domains.