Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 317
Filtrar
Mais filtros

Tipo de documento
Intervalo de ano de publicação
1.
J Neurosci ; 44(22)2024 May 29.
Artigo em Inglês | MEDLINE | ID: mdl-38589232

RESUMO

In developmental language disorder (DLD), learning to comprehend and express oneself with spoken language is impaired, but the reason for this remains unknown. Using millisecond-scale magnetoencephalography recordings combined with machine learning models, we investigated whether the possible neural basis of this disruption lies in poor cortical tracking of speech. The stimuli were common spoken Finnish words (e.g., dog, car, hammer) and sounds with corresponding meanings (e.g., dog bark, car engine, hammering). In both children with DLD (10 boys and 7 girls) and typically developing (TD) control children (14 boys and 3 girls), aged 10-15 years, the cortical activation to spoken words was best modeled as time-locked to the unfolding speech input at ∼100 ms latency between sound and cortical activation. Amplitude envelope (amplitude changes) and spectrogram (detailed time-varying spectral content) of the spoken words, but not other sounds, were very successfully decoded based on time-locked brain responses in bilateral temporal areas; based on the cortical responses, the models could tell at ∼75-85% accuracy which of the two sounds had been presented to the participant. However, the cortical representation of the amplitude envelope information was poorer in children with DLD compared with TD children at longer latencies (at ∼200-300 ms lag). We interpret this effect as reflecting poorer retention of acoustic-phonetic information in short-term memory. This impaired tracking could potentially affect the processing and learning of words as well as continuous speech. The present results offer an explanation for the problems in language comprehension and acquisition in DLD.


Assuntos
Transtornos do Desenvolvimento da Linguagem , Magnetoencefalografia , Percepção da Fala , Humanos , Masculino , Feminino , Criança , Adolescente , Magnetoencefalografia/métodos , Transtornos do Desenvolvimento da Linguagem/fisiopatologia , Percepção da Fala/fisiologia , Córtex Cerebral/fisiopatologia , Estimulação Acústica/métodos , Fala/fisiologia
2.
Cereb Cortex ; 34(1)2024 01 14.
Artigo em Inglês | MEDLINE | ID: mdl-38142293

RESUMO

Selective attention to one speaker in multi-talker environments can be affected by the acoustic and semantic properties of speech. One highly ecological feature of speech that has the potential to assist in selective attention is voice familiarity. Here, we tested how voice familiarity interacts with selective attention by measuring the neural speech-tracking response to both target and non-target speech in a dichotic listening "Cocktail Party" paradigm. We measured Magnetoencephalography from n = 33 participants, presented with concurrent narratives in two different voices, and instructed to pay attention to one ear ("target") and ignore the other ("non-target"). Participants were familiarized with one of the voices during the week prior to the experiment, rendering this voice familiar to them. Using multivariate speech-tracking analysis we estimated the neural responses to both stimuli and replicate their well-established modulation by selective attention. Importantly, speech-tracking was also affected by voice familiarity, showing enhanced response for target speech and reduced response for non-target speech in the contra-lateral hemisphere, when these were in a familiar vs. an unfamiliar voice. These findings offer valuable insight into how voice familiarity, and by extension, auditory-semantics, interact with goal-driven attention, and facilitate perceptual organization and speech processing in noisy environments.


Assuntos
Percepção da Fala , Voz , Humanos , Fala , Percepção da Fala/fisiologia , Reconhecimento Psicológico/fisiologia , Semântica
3.
Cereb Cortex ; 34(2)2024 01 31.
Artigo em Inglês | MEDLINE | ID: mdl-38265297

RESUMO

Numerous studies have been devoted to neural mechanisms of a variety of linguistic tasks (e.g. speech comprehension and production). To date, however, whether and how the neural patterns underlying different linguistic tasks are similar or differ remains elusive. In this study, we compared the neural patterns underlying 3 linguistic tasks mainly concerning speech comprehension and production. To address this, multivariate regression approaches with lesion/disconnection symptom mapping were applied to data from 216 stroke patients with damage to the left hemisphere. The results showed that lesion/disconnection patterns could predict both poststroke scores of speech comprehension and production tasks; these patterns exhibited shared regions on the temporal pole of the left hemisphere as well as unique regions contributing to the prediction for each domain. Lower scores in speech comprehension tasks were associated with lesions/abnormalities in the superior temporal gyrus and middle temporal gyrus, while lower scores in speech production tasks were associated with lesions/abnormalities in the left inferior parietal lobe and frontal lobe. These results suggested an important role of the ventral and dorsal stream pathways in speech comprehension and production (i.e. supporting the dual stream model) and highlighted the applicability of the novel multivariate disconnectome-based symptom mapping in cognitive neuroscience research.


Assuntos
Mapeamento Encefálico , Acidente Vascular Cerebral , Humanos , Mapeamento Encefálico/métodos , Imageamento por Ressonância Magnética/métodos , Linguística , Acidente Vascular Cerebral/complicações , Acidente Vascular Cerebral/diagnóstico por imagem , Compreensão
4.
J Neurosci ; 43(44): 7429-7440, 2023 11 01.
Artigo em Inglês | MEDLINE | ID: mdl-37793908

RESUMO

Selective attention to one of several competing speakers is required for comprehending a target speaker among other voices and for successful communication with them. It moreover has been found to involve the neural tracking of low-frequency speech rhythms in the auditory cortex. Effects of selective attention have also been found in subcortical neural activities, in particular regarding the frequency-following response related to the fundamental frequency of speech (speech-FFR). Recent investigations have, however, shown that the speech-FFR contains cortical contributions as well. It remains unclear whether these are also modulated by selective attention. Here we used magnetoencephalography to assess the attentional modulation of the cortical contributions to the speech-FFR. We presented both male and female participants with two competing speech signals and analyzed the cortical responses during attentional switching between the two speakers. Our findings revealed robust attentional modulation of the cortical contribution to the speech-FFR: the neural responses were higher when the speaker was attended than when they were ignored. We also found that, regardless of attention, a voice with a lower fundamental frequency elicited a larger cortical contribution to the speech-FFR than a voice with a higher fundamental frequency. Our results show that the attentional modulation of the speech-FFR does not only occur subcortically but extends to the auditory cortex as well.SIGNIFICANCE STATEMENT Understanding speech in noise requires attention to a target speaker. One of the speech features that a listener can use to identify a target voice among others and attend it is the fundamental frequency, together with its higher harmonics. The fundamental frequency arises from the opening and closing of the vocal folds and is tracked by high-frequency neural activity in the auditory brainstem and in the cortex. Previous investigations showed that the subcortical neural tracking is modulated by selective attention. Here we show that attention affects the cortical tracking of the fundamental frequency as well: it is stronger when a particular voice is attended than when it is ignored.


Assuntos
Córtex Auditivo , Percepção da Fala , Humanos , Masculino , Feminino , Fala , Percepção da Fala/fisiologia , Córtex Auditivo/fisiologia , Magnetoencefalografia , Potenciais Evocados Auditivos do Tronco Encefálico/fisiologia , Estimulação Acústica , Eletroencefalografia/métodos
5.
J Neurosci ; 43(32): 5856-5869, 2023 08 09.
Artigo em Inglês | MEDLINE | ID: mdl-37491313

RESUMO

Hearing impairment affects many older adults but is often diagnosed decades after speech comprehension in noisy situations has become effortful. Accurate assessment of listening effort may thus help diagnose hearing impairment earlier. However, pupillometry-the most used approach to assess listening effort-has limitations that hinder its use in practice. The current study explores a novel way to assess listening effort through eye movements. Building on cognitive and neurophysiological work, we examine the hypothesis that eye movements decrease when speech listening becomes challenging. In three experiments with human participants from both sexes, we demonstrate, consistent with this hypothesis, that fixation duration increases and spatial gaze dispersion decreases with increasing speech masking. Eye movements decreased during effortful speech listening for different visual scenes (free viewing, object tracking) and speech materials (simple sentences, naturalistic stories). In contrast, pupillometry was less sensitive to speech masking during story listening, suggesting pupillometric measures may not be as effective for the assessments of listening effort in naturalistic speech-listening paradigms. Our results reveal a critical link between eye movements and cognitive load, suggesting that neural activity in the brain regions that support the regulation of eye movements, such as frontal eye field and superior colliculus, are modulated when listening is effortful.SIGNIFICANCE STATEMENT Assessment of listening effort is critical for early diagnosis of age-related hearing loss. Pupillometry is most used but has several disadvantages. The current study explores a novel way to assess listening effort through eye movements. We examine the hypothesis that eye movements decrease when speech listening becomes effortful. We demonstrate, consistent with this hypothesis, that fixation duration increases and gaze dispersion decreases with increasing speech masking. Eye movements decreased during effortful speech listening for different visual scenes (free viewing, object tracking) and speech materials (sentences, naturalistic stories). Our results reveal a critical link between eye movements and cognitive load, suggesting that neural activity in brain regions that support the regulation of eye movements are modulated when listening is effortful.


Assuntos
Percepção da Fala , Fala , Masculino , Feminino , Humanos , Idoso , Movimentos Oculares , Percepção da Fala/fisiologia , Percepção Auditiva , Ruído , Inteligibilidade da Fala
6.
J Neurosci ; 43(27): 5045-5056, 2023 07 05.
Artigo em Inglês | MEDLINE | ID: mdl-37336758

RESUMO

The well-known "cocktail party effect" refers to incidental detection of salient words, such as one's own-name, in supposedly unattended speech. However, empirical investigation of the prevalence of this phenomenon and the underlying mechanisms has been limited to extremely artificial contexts and has yielded conflicting results. We introduce a novel empirical approach for revisiting this effect under highly ecological conditions, by immersing participants in a multisensory Virtual Café and using realistic stimuli and tasks. Participants (32 female, 18 male) listened to conversational speech from a character at their table, while a barista in the back of the café called out food orders. Unbeknownst to them, the barista sometimes called orders containing either their own-name or words that created semantic violations. We assessed the neurophysiological response-profile to these two probes in the task-irrelevant barista stream by measuring participants' brain activity (EEG), galvanic skin response and overt gaze-shifts.SIGNIFICANCE STATEMENT We found distinct neural and physiological responses to participants' own-name and semantic violations, indicating their incidental semantic processing despite being task-irrelevant. Interestingly, these responses were covert in nature and gaze-patterns were not associated with word-detection responses. This study emphasizes the nonexclusive nature of attention in multimodal ecological environments and demonstrates the brain's capacity to extract linguistic information from additional sources outside the primary focus of attention.


Assuntos
Semântica , Percepção da Fala , Humanos , Masculino , Feminino , Fala , Percepção Auditiva , Atenção/fisiologia , Linguística , Percepção da Fala/fisiologia
7.
J Neurosci ; 43(48): 8189-8200, 2023 11 29.
Artigo em Inglês | MEDLINE | ID: mdl-37793909

RESUMO

Spontaneous speech is produced in chunks called intonation units (IUs). IUs are defined by a set of prosodic cues and presumably occur in all human languages. Recent work has shown that across different grammatical and sociocultural conditions IUs form rhythms of ∼1 unit per second. Linguistic theory suggests that IUs pace the flow of information in the discourse. As a result, IUs provide a promising and hitherto unexplored theoretical framework for studying the neural mechanisms of communication. In this article, we identify a neural response unique to the boundary defined by the IU. We measured the EEG of human participants (of either sex), who listened to different speakers recounting an emotional life event. We analyzed the speech stimuli linguistically and modeled the EEG response at word offset using a GLM approach. We find that the EEG response to IU-final words differs from the response to IU-nonfinal words even when equating acoustic boundary strength. Finally, we relate our findings to the body of research on rhythmic brain mechanisms in speech processing. We study the unique contribution of IUs and acoustic boundary strength in predicting delta-band EEG. This analysis suggests that IU-related neural activity, which is tightly linked to the classic Closure Positive Shift (CPS), could be a time-locked component that captures the previously characterized delta-band neural speech tracking.SIGNIFICANCE STATEMENT Linguistic communication is central to human experience, and its neural underpinnings are a topic of much research in recent years. Neuroscientific research has benefited from studying human behavior in naturalistic settings, an endeavor that requires explicit models of complex behavior. Usage-based linguistic theory suggests that spoken language is prosodically structured in intonation units. We reveal that the neural system is attuned to intonation units by explicitly modeling their impact on the EEG response beyond mere acoustics. To our understanding, this is the first time this is demonstrated in spontaneous speech under naturalistic conditions and under a theoretical framework that connects the prosodic chunking of speech, on the one hand, with the flow of information during communication, on the other.


Assuntos
Percepção da Fala , Fala , Humanos , Fala/fisiologia , Eletroencefalografia , Estimulação Acústica , Percepção da Fala/fisiologia , Idioma
8.
J Neurosci ; 43(40): 6779-6795, 2023 10 04.
Artigo em Inglês | MEDLINE | ID: mdl-37607822

RESUMO

Communication difficulties are one of the core criteria in diagnosing autism spectrum disorder (ASD), and are often characterized by speech reception difficulties, whose biological underpinnings are not yet identified. This deficit could denote atypical neuronal ensemble activity, as reflected by neural oscillations. Atypical cross-frequency oscillation coupling, in particular, could disrupt the joint tracking and prediction of dynamic acoustic stimuli, a dual process that is essential for speech comprehension. Whether such oscillatory anomalies already exist in very young children with ASD, and with what specificity they relate to individual language reception capacity is unknown. We collected neural activity data using electroencephalography (EEG) in 64 very young children with and without ASD (mean age 3; 17 females, 47 males) while they were exposed to naturalistic-continuous speech. EEG power of frequency bands typically associated with phrase-level chunking (δ, 1-3 Hz), phonemic encoding (low-γ, 25-35 Hz), and top-down control (ß, 12-20 Hz) were markedly reduced in ASD relative to typically developing (TD) children. Speech neural tracking by δ and θ (4-8 Hz) oscillations was also weaker in ASD compared with TD children. After controlling gaze-pattern differences, we found that the classical θ/γ coupling was replaced by an atypical ß/γ coupling in children with ASD. This anomaly was the single most specific predictor of individual speech reception difficulties in ASD children. These findings suggest that early interventions (e.g., neurostimulation) targeting the disruption of ß/γ coupling and the upregulation of θ/γ coupling could improve speech processing coordination in young children with ASD and help them engage in oral interactions.SIGNIFICANCE STATEMENT Very young children already present marked alterations of neural oscillatory activity in response to natural speech at the time of autism spectrum disorder (ASD) diagnosis. Hierarchical processing of phonemic-range and syllabic-range information (θ/γ coupling) is disrupted in ASD children. Abnormal bottom-up (low-γ) and top-down (low-ß) coordination specifically predicts speech reception deficits in very young ASD children, and no other cognitive deficit.


Assuntos
Transtorno do Espectro Autista , Transtorno Autístico , Masculino , Feminino , Humanos , Criança , Pré-Escolar , Fala/fisiologia , Transtorno do Espectro Autista/diagnóstico , Eletroencefalografia , Estimulação Acústica
9.
Neuroimage ; 289: 120546, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38387743

RESUMO

The neuronal signatures of sensory and cognitive load provide access to brain activities related to complex listening situations. Sensory and cognitive loads are typically reflected in measures like response time (RT) and event-related potentials (ERPs) components. It's, however, strenuous to distinguish the underlying brain processes solely from these measures. In this study, along with RT- and ERP-analysis, we performed time-frequency analysis and source localization of oscillatory activity in participants performing two different auditory tasks with varying degrees of complexity and related them to sensory and cognitive load. We studied neuronal oscillatory activity in both periods before the behavioral response (pre-response) and after it (post-response). Robust oscillatory activities were found in both periods and were differentially affected by sensory and cognitive load. Oscillatory activity under sensory load was characterized by decrease in pre-response (early) theta activity and increased alpha activity. Oscillatory activity under cognitive load was characterized by increased theta activity, mainly in post-response (late) time. Furthermore, source localization revealed specific brain regions responsible for processing these loads, such as temporal and frontal lobe, cingulate cortex and precuneus. The results provide evidence that in complex listening situations, the brain processes sensory and cognitive loads differently. These neural processes have specific oscillatory signatures and are long lasting, extending beyond the behavioral response.


Assuntos
Eletroencefalografia , Potenciais Evocados , Humanos , Eletroencefalografia/métodos , Potenciais Evocados/fisiologia , Encéfalo/fisiologia , Lobo Frontal , Cognição/fisiologia
10.
Hum Brain Mapp ; 45(8): e26676, 2024 Jun 01.
Artigo em Inglês | MEDLINE | ID: mdl-38798131

RESUMO

Aphasia is a communication disorder that affects processing of language at different levels (e.g., acoustic, phonological, semantic). Recording brain activity via Electroencephalography while people listen to a continuous story allows to analyze brain responses to acoustic and linguistic properties of speech. When the neural activity aligns with these speech properties, it is referred to as neural tracking. Even though measuring neural tracking of speech may present an interesting approach to studying aphasia in an ecologically valid way, it has not yet been investigated in individuals with stroke-induced aphasia. Here, we explored processing of acoustic and linguistic speech representations in individuals with aphasia in the chronic phase after stroke and age-matched healthy controls. We found decreased neural tracking of acoustic speech representations (envelope and envelope onsets) in individuals with aphasia. In addition, word surprisal displayed decreased amplitudes in individuals with aphasia around 195 ms over frontal electrodes, although this effect was not corrected for multiple comparisons. These results show that there is potential to capture language processing impairments in individuals with aphasia by measuring neural tracking of continuous speech. However, more research is needed to validate these results. Nonetheless, this exploratory study shows that neural tracking of naturalistic, continuous speech presents a powerful approach to studying aphasia.


Assuntos
Afasia , Eletroencefalografia , Acidente Vascular Cerebral , Humanos , Afasia/fisiopatologia , Afasia/etiologia , Afasia/diagnóstico por imagem , Masculino , Feminino , Pessoa de Meia-Idade , Acidente Vascular Cerebral/complicações , Acidente Vascular Cerebral/fisiopatologia , Idoso , Percepção da Fala/fisiologia , Adulto , Fala/fisiologia
11.
Hum Brain Mapp ; 45(1): e26577, 2024 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-38224542

RESUMO

Healthy aging leads to complex changes in the functional network of speech processing in a noisy environment. The dual-route neural architecture has been applied to the study of speech processing. Although evidence suggests that senescent increases activity in the brain regions across the dorsal and ventral stream regions to offset reduced periphery, the regulatory mechanism of dual-route functional networks underlying such compensation remains largely unknown. Here, by utilizing functional near-infrared spectroscopy (fNIRS), we investigated the compensatory mechanism of the dual-route functional connectivity, and its relationship with healthy aging by using a speech perception task at varying signal-to-noise ratios (SNR) in healthy individuals (young adults, middle-aged adults, and older adults). Results showed that the speech perception scores showed a significant age-related decrease with the reduction of the SNR. The analysis results of dual-route speech processing networks showed that the functional connection of Wernicke's area and homolog Wernicke's area were age-related increases. Further to clarify the age-related characteristics of the dual-route speech processing networks, graph-theoretical network analysis revealed an age-related increase in the efficiency of the networks, and the age-related differences in nodal characteristics were found both in Wernicke's area and homolog Wernicke's area under noise environment. Thus, Wernicke's area might be a key network hub to maintain efficient information transfer across the speech process network with healthy aging. Moreover, older adults would recruit more resources from the homologous Wernicke's area in a noisy environment. The recruitment of the homolog of Wernicke's area might provide a means of compensation for older adults for decoding speech in an adverse listening environment. Together, our results characterized dual-route speech processing networks at varying noise environments and provided new insight for the compensatory theories of how aging modulates the dual-route speech processing functional networks.


Assuntos
Percepção da Fala , Fala , Pessoa de Meia-Idade , Adulto Jovem , Humanos , Idoso , Imageamento por Ressonância Magnética , Envelhecimento , Encéfalo/diagnóstico por imagem
12.
Brain Cogn ; 174: 106120, 2024 02.
Artigo em Inglês | MEDLINE | ID: mdl-38142535

RESUMO

Previous studies found that prolonged musical training can promote language processing, but few studies have examined whether and how musical training affects the processing of accentuation in spoken language. In this study, a vocabulary detection task was conducted, with Chinese single sentences as materials, to investigate how musicians and non-musicians process corrective accent and information accent in the sentence-middle and sentence-final positions. In the sentence-middle position, results of the cluster-based permutation t-tests showed significant differences in the 574-714 ms time window for the control group. In the sentence-final position, the cluster-based permutation t-tests revealed significant differences in the 612-810 ms time window for the music group and in the 616-812 ms time window for the control group. These significant positive effects were induced by the processing of information accent relative to that of corrective accent. These results suggest that both groups were able to distinguish corrective accent from information accent, but they processed the two accent types differently in the sentence-middle position. These findings show that musical training has a cross-domain effect on spoken language processing and that the accent position also affects its processing.


Assuntos
Música , Percepção da Fala , Humanos , Idioma , Potenciais Evocados , Vocabulário
13.
Cereb Cortex ; 33(9): 5361-5374, 2023 04 25.
Artigo em Inglês | MEDLINE | ID: mdl-36331339

RESUMO

Many situations require focusing attention on one speaker, while monitoring the environment for potentially important information. Some have proposed that dividing attention among 2 speakers involves behavioral trade-offs, due to limited cognitive resources. However the severity of these trade-offs, particularly under ecologically-valid circumstances, is not well understood. We investigated the capacity to process simultaneous speech using a dual-task paradigm simulating task-demands and stimuli encountered in real-life. Participants listened to conversational narratives (Narrative Stream) and monitored a stream of announcements (Barista Stream), to detect when their order was called. We measured participants' performance, neural activity, and skin conductance as they engaged in this dual-task. Participants achieved extremely high dual-task accuracy, with no apparent behavioral trade-offs. Moreover, robust neural and physiological responses were observed for target-stimuli in the Barista Stream, alongside significant neural speech-tracking of the Narrative Stream. These results suggest that humans have substantial capacity to process simultaneous speech and do not suffer from insufficient processing resources, at least for this highly ecological task-combination and level of perceptual load. Results also confirmed the ecological validity of the advantage for detecting ones' own name at the behavioral, neural, and physiological level, highlighting the contribution of personal relevance when processing simultaneous speech.


Assuntos
Percepção da Fala , Fala , Humanos , Fala/fisiologia , Percepção da Fala/fisiologia , Percepção Auditiva , Atenção/fisiologia
14.
Cereb Cortex ; 33(18): 10036-10046, 2023 09 09.
Artigo em Inglês | MEDLINE | ID: mdl-37491998

RESUMO

Speech comprehension is a complex process involving multiple stages, such as decoding of phonetic units, recognizing words, and understanding sentences and passages. In this study, we identify cortical networks beyond basic phonetic processing using a novel passage learning paradigm. Participants learn to comprehend a story composed of syllables of their native language, but containing unfamiliar vocabulary and syntax. Three learning methods are employed, each resulting in some degree of learning within a 12-min learning session. Functional magnetic resonance imaging results reveal that, when listening to the same story, the classic temporal-frontal language network is significantly enhanced by learning. Critically, activation of the left anterior and posterior temporal lobe correlates with the learning outcome that is assessed behaviorally through, e.g. word recognition and passage comprehension tests. This study demonstrates that a brief learning session is sufficient to induce neural plasticity in the left temporal lobe, which underlies the transformation from phonetic units to the units of meaning, such as words and sentences.


Assuntos
Percepção da Fala , Vocabulário , Humanos , Aprendizagem , Idioma , Fala , Fonética , Percepção da Fala/fisiologia , Imageamento por Ressonância Magnética/métodos , Compreensão/fisiologia , Mapeamento Encefálico
15.
Cogn Emot ; : 1-10, 2024 May 19.
Artigo em Inglês | MEDLINE | ID: mdl-38764186

RESUMO

Older adults process emotional speech differently than young adults, relying less on prosody (tone) relative to semantics (words). This study aimed to elucidate the mechanisms underlying these age-related differences via an emotional speech-in-noise test. A sample of 51 young and 47 older adults rated spoken sentences with emotional content on both prosody and semantics, presented on the background of wideband speech-spectrum noise (sensory interference) or on the background of multi-talker babble (sensory/cognitive interference). The presence of wideband noise eliminated age-related differences in semantics but not in prosody when processing emotional speech. Conversely, the presence of babble resulted in the elimination of age-related differences across all measures. The results suggest that both sensory and cognitive-linguistic factors contribute to age-related changes in emotional speech processing. Because real world conditions typically involve noisy background, our results highlight the importance of testing under such conditions.

16.
Artigo em Inglês | MEDLINE | ID: mdl-38726473

RESUMO

BACKGROUND: Cleft lip and palate is one of the most common oral and maxillofacial deformities associated with a variety of functional disorders. Cleft palate speech disorder (CPSD) occurs the most frequently and manifests a series of characteristic speech features, which are called cleft speech characteristics. Some scholars believe that children with CPSD and poor speech outcomes may also have weaknesses in speech input processing ability, but evidence is still lacking so far. AIMS: (1) To explore whether children with CPSD and speech output disorders also have defects in speech input processing abilities; (2) to explore the correlation between speech input and output processing abilities. METHODS & PROCEDURES: Children in the experimental group were enrolled from Beijing Stomatological Hospital, Capital Medical University, and healthy volunteers were recruited as controls. Then three tasks containing real and pseudo words were performed sequentially. Reaction time, accuracy and other indicators in three tasks were collected and then analysed. OUTCOMES & RESULTS: The indicators in the experimental group were significantly lower than those in the control group. There was a strong correlation between speech input and output processing tasks. The performance of both groups when processing pseudo words in the three tasks was worse than that when dealing with real words. CONCLUSIONS & IMPLICATIONS: Compared with normal controls, children with CPSD have deficits in both speech input and output processing, and there is a strong correlation between speech input and output speech processing abilities. In addition, the pseudo words task was more challenging than the real word task for both groups. WHAT THIS PAPER ADDS: What is already known on the subject Children with cleft lip and palate often have speech sound disorders known as cleft palate speech disorder (CPSD). CPSD is characterised by consonant errors called cleft speech characteristics, which can persist even after surgery. Some studies suggest that poor speech outcomes in children with CPSD may be associated with deficits in processing speech input. However, this has not been validated in mainland China. What this paper adds to existing knowledge The results of our study indicate that children with CPSD exhibit poorer performance in three tasks assessing speech input and output abilities compared to healthy controls, suggesting their deficits in both speech input and output processing. Furthermore, a significant correlation was observed between speech input and output processing abilities. Additionally, both groups demonstrated greater difficulty in processing pseudo words compared to real words, as evidenced by their worse performance in dealing with pseudo words. What are the potential or actual clinical implications of this work? The pseudo word tasks designed and implemented in our study can be employed in future research and assessment of speech input and output abilities in Chinese Mandarin children with CPSD. Additionally, our findings revealed the significance of considering both speech output processing abilities and potential existence of speech input processing ability for speech and language therapists when evaluating and developing treatment options for children with CPSD as these abilities are also important for the development of literacy development.

17.
Artigo em Inglês | MEDLINE | ID: mdl-38206308

RESUMO

BACKGROUND: Classification systems in healthcare support shared understanding of conditions for clinical communication, service monitoring and development, and research. Children born with cleft palate with or without cleft lip (CP+/-L) are at high risk of developing cleft-related speech sound disorder (SSD). The way cleft-related SSD is represented and described in SSD classification systems varies. Reflecting on the potential causal pathways for different cleft-related speech features, including the role of speech processing skills, may inform how cleft-related SSD is represented in classification systems. AIM & APPROACH: To explore and reflect on how cleft-related SSD is represented in current SSD classification systems in the context of considering how speech processing skills and other factors may be involved in causal pathways of cleft speech characteristics (CSCs). MAIN CONTRIBUTION: Variation in the representation of cleft-related SSD in classification systems is described. Potential causal pathways for passive cleft- related speech features and different active CSCs are explored. The factors involved in the development and/or persistence of different active CSCs may vary. Some factors may be specific to children born with CP+/-L, but if speech processing skills are also involved, this is an overlap with other SSD subtypes. Current evidence regarding relationships between different speech processing skills and active CSCs is limited. Implications for the representation of cleft-related SSD in SSD classification systems are discussed. CONCLUSION: There are different categories of cleft-related speech features which are essential to understand and identify in children with cleft-related SSD to ensure appropriate management. Representation of these feature categories in classification systems could support understanding of speech in this population. Speech processing skills could be involved in the development and/or persistence of different active CSCs in individual children. Reflection and discussion on how cleft-related SSD is represented in classification systems in relation to other SSD subtypes may inform future iterations of these systems. Further work is needed to understand factors influencing the development and/or persistence of active CSCs, including speech processing skills. WHAT THIS PAPER ADDS: What is already known on the subject Cleft-related speech sound disorder (SSD) is commonly described as being of known origin. The features of cleft-related SSD have been described extensively and several authors have also examined factors which may contribute to speech development and outcomes in children born with cleft palate +/- lip. There is limited evidence regarding the role of speech processing in the development and persistence of cleft-related SSD. What this study adds This paper reflects on how cleft-related SSD is represented in SSD classification systems in relation to key feature categories of cleft-related SSD and possible causal pathways for passive features and active cleft speech characteristics (CSCs). The role of cognitive speech processing skills is specifically considered alongside other factors that may contribute to the development of active CSCs. What are the clinical implications of this work? Causal pathways for different features of cleft-related SSD may vary, particularly between passive and active features, abut also between different active CSCs. Speech and language therapists (SLTs) need to differentially diagnose passive speech features and active CSCs. Consideration of the role of different speech processing skills and interactions with other potentially influencing factors in relation to active CSCs may inform clinical hypotheses and speech and language therapy (SLT) intervention. Representing key features of cleft-related SSD in classification systems may support understanding of cleft-related SSD in relation to other SSD subtypes.

18.
Int J Lang Commun Disord ; 59(4): 1422-1435, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38237606

RESUMO

BACKGROUND: Perceptual measures such as speech intelligibility are known to be biased, variant and subjective, to which an automatic approach has been seen as a more reliable alternative. On the other hand, automatic approaches tend to lack explainability, an aspect that can prevent the widespread usage of these technologies clinically. AIMS: In the present work, we aim to study the relationship between four perceptual parameters and speech intelligibility by automatically modelling the behaviour of six perceptual judges, in the context of head and neck cancer. From this evaluation we want to assess the different levels of relevance of each parameter as well as the different judge profiles that arise, both perceptually and automatically. METHODS AND PROCEDURES: Based on a passage reading task from the Carcinologic Speech Severity Index (C2SI) corpus, six expert listeners assessed the voice quality, resonance, prosody and phonemic distortions, as well as the speech intelligibility of patients treated for oral or oropharyngeal cancer. A statistical analysis and an ensemble of automatic systems, one per judge, were devised, where speech intelligibility is predicted as a function of the four aforementioned perceptual parameters of voice quality, resonance, prosody and phonemic distortions. OUTCOMES AND RESULTS: The results suggest that we can automatically predict speech intelligibility as a function of the four aforementioned perceptual parameters, achieving a high correlation of 0.775 (Spearman's ρ). Furthermore, different judge profiles were found perceptually that were successfully modelled automatically. CONCLUSIONS AND IMPLICATIONS: The four investigated perceptual parameters influence the global rating of speech intelligibility, showing that different judge profiles emerge. The proposed automatic approach displayed a more uniform profile across all judges, displaying a more reliable, unbiased and objective prediction. The system also adds an extra layer of interpretability, since speech intelligibility is regressed as a direct function of the individual prediction of the four perceptual parameters, an improvement over more black box approaches. WHAT THIS PAPER ADDS: What is already known on this subject Speech intelligibility is a clinical measure typically used in the post-treatment assessment of speech affecting disorders, such as head and neck cancer. Their perceptual assessment is currently the main method of evaluation; however, it is known to be quite subjective since intelligibility can be seen as a combination of other perceptual parameters (voice quality, resonance, etc.). Given this, automatic approaches have been seen as a more viable alternative to the traditionally used perceptual assessments. What this study adds to existing knowledge The present work introduces a study based on the relationship between four perceptual parameters (voice quality, resonance, prosody and phonemic distortions) and speech intelligibility, by automatically modelling the behaviour of six perceptual judges. The results suggest that different judge profiles arise, both in the perceptual case as well as in the automatic models. These different profiles found showcase the different schools of thought that perceptual judges have, in comparison to the automatic judges, that display more uniform levels of relevance across all the four perceptual parameters. This aspect shows that an automatic approach promotes unbiased, reliable and more objective predictions. What are the clinical implications of this work? The automatic prediction of speech intelligibility, using a combination of four perceptual parameters, show that these approaches can achieve high correlations with the reference scores while maintaining a certain degree of explainability. The more uniform judge profiles found on the automatic case also display less biased results towards the four perceptual parameters. This aspect facilitates the clinical implementation of this class of systems, as opposed to the more subjective and harder to reproduce perceptual assessments.


Assuntos
Neoplasias de Cabeça e Pescoço , Inteligibilidade da Fala , Humanos , Masculino , Feminino , Neoplasias de Cabeça e Pescoço/psicologia , Pessoa de Meia-Idade , Idoso , Julgamento , Percepção da Fala , Qualidade da Voz , Adulto
19.
Neuroimage ; 267: 119841, 2023 02 15.
Artigo em Inglês | MEDLINE | ID: mdl-36584758

RESUMO

BACKGROUND: Older adults process speech differently, but it is not yet clear how aging affects different levels of processing natural, continuous speech, both in terms of bottom-up acoustic analysis and top-down generation of linguistic-based predictions. We studied natural speech processing across the adult lifespan via electroencephalography (EEG) measurements of neural tracking. GOALS: Our goals are to analyze the unique contribution of linguistic speech processing across the adult lifespan using natural speech, while controlling for the influence of acoustic processing. Moreover, we also studied acoustic processing across age. In particular, we focus on changes in spatial and temporal activation patterns in response to natural speech across the lifespan. METHODS: 52 normal-hearing adults between 17 and 82 years of age listened to a naturally spoken story while the EEG signal was recorded. We investigated the effect of age on acoustic and linguistic processing of speech. Because age correlated with hearing capacity and measures of cognition, we investigated whether the observed age effect is mediated by these factors. Furthermore, we investigated whether there is an effect of age on hemisphere lateralization and on spatiotemporal patterns of the neural responses. RESULTS: Our EEG results showed that linguistic speech processing declines with advancing age. Moreover, as age increased, the neural response latency to certain aspects of linguistic speech processing increased. Also acoustic neural tracking (NT) decreased with increasing age, which is at odds with the literature. In contrast to linguistic processing, older subjects showed shorter latencies for early acoustic responses to speech. No evidence was found for hemispheric lateralization in neither younger nor older adults during linguistic speech processing. Most of the observed aging effects on acoustic and linguistic processing were not explained by age-related decline in hearing capacity or cognition. However, our results suggest that the effect of decreasing linguistic neural tracking with advancing age at word-level is also partially due to an age-related decline in cognition than a robust effect of age. CONCLUSION: Spatial and temporal characteristics of the neural responses to continuous speech change across the adult lifespan for both acoustic and linguistic speech processing. These changes may be traces of structural and/or functional change that occurs with advancing age.


Assuntos
Percepção da Fala , Fala , Humanos , Idoso , Fala/fisiologia , Estimulação Acústica/métodos , Percepção da Fala/fisiologia , Eletroencefalografia/métodos , Linguística , Acústica
20.
Neuroimage ; 278: 120285, 2023 09.
Artigo em Inglês | MEDLINE | ID: mdl-37481009

RESUMO

Healthy aging is often associated with speech comprehension difficulties in everyday life situations despite a pure-tone hearing threshold in the normative range. Drawing on this background, we used a multidimensional approach to assess the functional and structural neural correlates underlying age-related temporal speech processing while controlling for pure-tone hearing acuity. Accordingly, we combined structural magnetic resonance imaging and electroencephalography, and collected behavioral data while younger and older adults completed a phonetic categorization and discrimination task with consonant-vowel syllables varying along a voice-onset time continuum. The behavioral results confirmed age-related temporal speech processing singularities which were reflected in a shift of the boundary of the psychometric categorization function, with older adults perceiving more syllable characterized by a short voice-onset time as /ta/ compared to younger adults. Furthermore, despite the absence of any between-group differences in phonetic discrimination abilities, older adults demonstrated longer N100/P200 latencies as well as increased P200 amplitudes while processing the consonant-vowel syllables varying in voice-onset time. Finally, older adults also exhibited a divergent anatomical gray matter infrastructure in bilateral auditory-related and frontal brain regions, as manifested in reduced cortical thickness and surface area. Notably, in the younger adults but not in the older adult cohort, cortical surface area in these two gross anatomical clusters correlated with the categorization of consonant-vowel syllables characterized by a short voice-onset time, suggesting the existence of a critical gray matter threshold that is crucial for consistent mapping of phonetic categories varying along the temporal dimension. Taken together, our results highlight the multifaceted dimensions of age-related temporal speech processing characteristics, and pave the way toward a better understanding of the relationships between hearing, speech and the brain in older age.


Assuntos
Percepção da Fala , Fala , Humanos , Idoso , Estimulação Acústica , Percepção Auditiva , Eletroencefalografia
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA