Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 823
Filtrar
1.
J Int Adv Otol ; 20(4): 289-300, 2024 Jul 29.
Artigo em Inglês | MEDLINE | ID: mdl-39159037

RESUMO

People with single-sided deafness (SSD) or asymmetric hearing loss (AHL) have particular difficulty understanding speech in noisy listening situations and in sound localization. The objective of this multicenter study is to evaluate the effect of a cochlear implant (CI) in adults with single-sided deafness (SSD) or asymmetric hearing loss (AHL), particularly regarding sound localization and speech intelligibility with additional interest in electric-acoustic pitch matching. A prospective longitudinal study at 7 European tertiary referral centers was conducted including 19 SSD and 16 AHL subjects undergoing cochlear implantation. Sound localization accuracy was investigated in terms of root mean square error and signed bias before and after implantation. Speech recognition in quiet and speech reception thresholds in noise for several spatial configurations were assessed preoperatively and at several post-activation time points. Pitch perception with CI was tracked using pitch matching. Data up to 12 months post activation were collected. In both SSD and AHL subjects, CI significantly improved sound localization for sound sources on the implant side, and thus overall sound localization. Speech recognition in quiet with the implant ear improved significantly. In noise, a significant head shadow effect was found for SSD subjects only. However, the evaluation of AHL subjects was limited by the small sample size. No uniform development of pitch perception with the implant ear was observed. The benefits shown in this study confirm and expand the existing body of evidence for the effectiveness of CI in SSD and AHL. Particularly, improved localization was shown to result from increased localization accuracy on the implant side.


Assuntos
Implante Coclear , Implantes Cocleares , Perda Auditiva Unilateral , Localização de Som , Percepção da Fala , Humanos , Implante Coclear/métodos , Masculino , Localização de Som/fisiologia , Feminino , Pessoa de Meia-Idade , Percepção da Fala/fisiologia , Estudos Prospectivos , Perda Auditiva Unilateral/cirurgia , Perda Auditiva Unilateral/reabilitação , Perda Auditiva Unilateral/fisiopatologia , Seguimentos , Idoso , Adulto , Europa (Continente) , Estudos Longitudinais , Resultado do Tratamento , Inteligibilidade da Fala/fisiologia , Percepção da Altura Sonora/fisiologia , Surdez/cirurgia , Surdez/reabilitação , Surdez/fisiopatologia , Ruído
2.
Proc Natl Acad Sci U S A ; 121(34): e2411167121, 2024 Aug 20.
Artigo em Inglês | MEDLINE | ID: mdl-39136991

RESUMO

Evidence accumulates that the cerebellum's role in the brain is not restricted to motor functions. Rather, cerebellar activity seems to be crucial for a variety of tasks that rely on precise event timing and prediction. Due to its complex structure and importance in communication, human speech requires a particularly precise and predictive coordination of neural processes to be successfully comprehended. Recent studies proposed that the cerebellum is indeed a major contributor to speech processing, but how this contribution is achieved mechanistically remains poorly understood. The current study aimed to reveal a mechanism underlying cortico-cerebellar coordination and demonstrate its speech-specificity. In a reanalysis of magnetoencephalography data, we found that activity in the cerebellum aligned to rhythmic sequences of noise-vocoded speech, irrespective of its intelligibility. We then tested whether these "entrained" responses persist, and how they interact with other brain regions, when a rhythmic stimulus stopped and temporal predictions had to be updated. We found that only intelligible speech produced sustained rhythmic responses in the cerebellum. During this "entrainment echo," but not during rhythmic speech itself, cerebellar activity was coupled with that in the left inferior frontal gyrus, and specifically at rates corresponding to the preceding stimulus rhythm. This finding represents evidence for specific cerebellum-driven temporal predictions in speech processing and their relay to cortical regions.


Assuntos
Cerebelo , Magnetoencefalografia , Humanos , Cerebelo/fisiologia , Masculino , Feminino , Adulto , Percepção da Fala/fisiologia , Adulto Jovem , Fala/fisiologia , Inteligibilidade da Fala/fisiologia
3.
Trends Hear ; 28: 23312165241240572, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38676325

RESUMO

Realistic outcome measures that reflect everyday hearing challenges are needed to assess hearing aid and cochlear implant (CI) fitting. Literature suggests that listening effort measures may be more sensitive to differences between hearing-device settings than established speech intelligibility measures when speech intelligibility is near maximum. Which method provides the most effective measurement of listening effort for this purpose is currently unclear. This study aimed to investigate the feasibility of two tests for measuring changes in listening effort in CI users due to signal-to-noise ratio (SNR) differences, as would arise from different hearing-device settings. By comparing the effect size of SNR differences on listening effort measures with test-retest differences, the study evaluated the suitability of these tests for clinical use. Nineteen CI users underwent two listening effort tests at two SNRs (+4 and +8 dB relative to individuals' 50% speech perception threshold). We employed dual-task paradigms-a sentence-final word identification and recall test (SWIRT) and a sentence verification test (SVT)-to assess listening effort at these two SNRs. Our results show a significant difference in listening effort between the SNRs for both test methods, although the effect size was comparable to the test-retest difference, and the sensitivity was not superior to speech intelligibility measures. Thus, the implementations of SVT and SWIRT used in this study are not suitable for clinical use to measure listening effort differences of this magnitude in individual CI users. However, they can be used in research involving CI users to analyze group data.


Assuntos
Implante Coclear , Implantes Cocleares , Estudos de Viabilidade , Pessoas com Deficiência Auditiva , Inteligibilidade da Fala , Percepção da Fala , Humanos , Masculino , Feminino , Percepção da Fala/fisiologia , Pessoa de Meia-Idade , Idoso , Inteligibilidade da Fala/fisiologia , Implante Coclear/instrumentação , Pessoas com Deficiência Auditiva/reabilitação , Pessoas com Deficiência Auditiva/psicologia , Reprodutibilidade dos Testes , Estimulação Acústica , Razão Sinal-Ruído , Adulto , Idoso de 80 Anos ou mais , Limiar Auditivo/fisiologia , Valor Preditivo dos Testes , Correção de Deficiência Auditiva/instrumentação , Ruído/efeitos adversos
4.
J Speech Lang Hear Res ; 67(4): 1090-1106, 2024 Apr 08.
Artigo em Inglês | MEDLINE | ID: mdl-38498664

RESUMO

PURPOSE: This study examined speech changes induced by deep-brain stimulation (DBS) in speakers with Parkinson's disease (PD) using a set of auditory-perceptual and acoustic measures. METHOD: Speech recordings from nine speakers with PD and DBS were compared between DBS-On and DBS-Off conditions using auditory-perceptual and acoustic analyses. Auditory-perceptual ratings included voice quality, articulation precision, prosody, speech intelligibility, and listening effort obtained from 44 listeners. Acoustic measures were made for voicing proportion, second formant frequency slope, vowel dispersion, articulation rate, and range of fundamental frequency and intensity. RESULTS: No significant changes were found between DBS-On and DBS-Off for the five perceptual ratings. Four of six acoustic measures revealed significant differences between the two conditions. While articulation rate and acoustic vowel dispersion increased, voicing proportion and intensity range decreased from the DBS-Off to DBS-On condition. However, a visual examination of the data indicated that the statistical significance was mostly driven by a small number of participants, while the majority did not show a consistent pattern of such changes. CONCLUSIONS: Our data, in general, indicate no-to-minimal changes in speech production ensued from DBS stimulation. The findings are discussed with a focus on large interspeaker variability in PD in terms of their speech characteristics and the potential effects of DBS on speech.


Assuntos
Estimulação Encefálica Profunda , Doença de Parkinson , Humanos , Acústica , Inteligibilidade da Fala/fisiologia , Qualidade da Voz , Doença de Parkinson/complicações , Doença de Parkinson/terapia , Encéfalo , Acústica da Fala
5.
Otol Neurotol ; 45(5): e385-e392, 2024 Jun 01.
Artigo em Inglês | MEDLINE | ID: mdl-38518764

RESUMO

HYPOTHESIS: The behaviorally based programming with loudness balancing (LB) would result in better speech understanding, spectral-temporal resolution, and music perception scores, and there would be a relationship between these scores. BACKGROUND: Loudness imbalances at upper stimulation levels may cause sounds to be perceived as irregular, gravelly, or overly echoed and may negatively affect the listening performance of the cochlear implant (CI) user. LB should be performed after fitting to overcome these problems. METHODS: The study included 26 unilateral Med-EL CI users. Two different CI programs based on the objective electrically evoked stapedial reflex threshold (P1) and the behaviorally program with LB (P2) were recorded for each participant. The Turkish Matrix Sentence Test (TMS) was applied to evaluate speech perception; the Random Gap Detection Test (RGDT) and Spectral-Temporally Modulated Ripple Test (SMRT) were applied to evaluate spectral temporal resolution skills; the Mini Profile of Music Perception Skills (mini-PROMS) and Melodic Contour Identification (MCI) tests were applied to evaluate music perception, and the results were compared. RESULTS: Significantly better scores were obtained with P2 in TMS tests performed in noise and quiet. SMRT scores were significantly correlated with TMS in quiet and noise, and mini-PROMS sound perception results. Although better scores were obtained with P2 in the mini-PROMS total score and MCI, a significant difference was found only for MCI. CONCLUSION: The data from the current study showed that equalization of loudness across CI electrodes leads to better perceptual acuity. It also revealed the relationship between speech perception, spectral-temporal resolution, and music perception.


Assuntos
Implante Coclear , Implantes Cocleares , Música , Percepção da Fala , Humanos , Masculino , Feminino , Pessoa de Meia-Idade , Adulto , Percepção da Fala/fisiologia , Implante Coclear/métodos , Inteligibilidade da Fala/fisiologia , Idoso , Percepção Auditiva/fisiologia , Percepção Sonora/fisiologia , Adulto Jovem
6.
Trends Hear ; 28: 23312165241232551, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38549351

RESUMO

In daily life, both acoustic factors and social context can affect listening effort investment. In laboratory settings, information about listening effort has been deduced from pupil and cardiovascular responses independently. The extent to which these measures can jointly predict listening-related factors is unknown. Here we combined pupil and cardiovascular features to predict acoustic and contextual aspects of speech perception. Data were collected from 29 adults (mean  =  64.6 years, SD  =  9.2) with hearing loss. Participants performed a speech perception task at two individualized signal-to-noise ratios (corresponding to 50% and 80% of sentences correct) and in two social contexts (the presence and absence of two observers). Seven features were extracted per trial: baseline pupil size, peak pupil dilation, mean pupil dilation, interbeat interval, blood volume pulse amplitude, pre-ejection period and pulse arrival time. These features were used to train k-nearest neighbor classifiers to predict task demand, social context and sentence accuracy. The k-fold cross validation on the group-level data revealed above-chance classification accuracies: task demand, 64.4%; social context, 78.3%; and sentence accuracy, 55.1%. However, classification accuracies diminished when the classifiers were trained and tested on data from different participants. Individually trained classifiers (one per participant) performed better than group-level classifiers: 71.7% (SD  =  10.2) for task demand, 88.0% (SD  =  7.5) for social context, and 60.0% (SD  =  13.1) for sentence accuracy. We demonstrated that classifiers trained on group-level physiological data to predict aspects of speech perception generalized poorly to novel participants. Individually calibrated classifiers hold more promise for future applications.


Assuntos
Pupila , Percepção da Fala , Humanos , Pupila/fisiologia , Inteligibilidade da Fala/fisiologia , Percepção da Fala/fisiologia , Pessoa de Meia-Idade , Idoso
7.
JASA Express Lett ; 4(2)2024 Feb 01.
Artigo em Inglês | MEDLINE | ID: mdl-38350077

RESUMO

Measuring how well human listeners recognize speech under varying environmental conditions (speech intelligibility) is a challenge for theoretical, technological, and clinical approaches to speech communication. The current gold standard-human transcription-is time- and resource-intensive. Recent advances in automatic speech recognition (ASR) systems raise the possibility of automating intelligibility measurement. This study tested 4 state-of-the-art ASR systems with second language speech-in-noise and found that one, whisper, performed at or above human listener accuracy. However, the content of whisper's responses diverged substantially from human responses, especially at lower signal-to-noise ratios, suggesting both opportunities and limitations for ASR--based speech intelligibility modeling.


Assuntos
Percepção da Fala , Humanos , Percepção da Fala/fisiologia , Ruído/efeitos adversos , Inteligibilidade da Fala/fisiologia , Interface para o Reconhecimento da Fala , Reconhecimento Psicológico
8.
PLoS Biol ; 22(2): e3002498, 2024 Feb.
Artigo em Inglês | MEDLINE | ID: mdl-38358954

RESUMO

Speech recognition crucially relies on slow temporal modulations (<16 Hz) in speech. Recent studies, however, have demonstrated that the long-delay echoes, which are common during online conferencing, can eliminate crucial temporal modulations in speech but do not affect speech intelligibility. Here, we investigated the underlying neural mechanisms. MEG experiments demonstrated that cortical activity can effectively track the temporal modulations eliminated by an echo, which cannot be fully explained by basic neural adaptation mechanisms. Furthermore, cortical responses to echoic speech can be better explained by a model that segregates speech from its echo than by a model that encodes echoic speech as a whole. The speech segregation effect was observed even when attention was diverted but would disappear when segregation cues, i.e., speech fine structure, were removed. These results strongly suggested that, through mechanisms such as stream segregation, the auditory system can build an echo-insensitive representation of speech envelope, which can support reliable speech recognition.


Assuntos
Córtex Auditivo , Percepção da Fala , Humanos , Percepção da Fala/fisiologia , Inteligibilidade da Fala/fisiologia , Encéfalo , Córtex Auditivo/fisiologia , Atenção , Estimulação Acústica
9.
Eur Arch Otorhinolaryngol ; 281(7): 3461-3473, 2024 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-38219245

RESUMO

PURPOSE: The purpose of this retrospective study is to compare the results of electrically evoked compound action potential (ECAP) measurements using automatic auditory response telemetry (AutoART) with those obtained by ART in adults. The study also aimed to evaluate the predictive value of intraoperative ART and AutoART ECAPs for speech intelligibility (SI) and hearing success (HS), and to determine if cochlear nerve (CN) cross-sectional area (CSA) obtained preoperatively by magnetic resonance imaging (MRI) scans could predict ART and AutoART ECAPs and SI and HS outcome. METHODS: The study analyzed and correlated ART and AutoART ECAP thresholds at electrodes E2, E6, and E10, as well as averaged ECAP thresholds over electrodes E1-E12, using data from 32 implants. Correlations were also examined for ART and AutoART ECAP slopes. In addition, averaged ART and AutoART ECAP thresholds and slopes over all 12 electrodes for each participant were correlated with CN CSA measured from MRI sequences. SI of the monosyllabic Freiburg Speech Test at 65 dB sound pressure level was examined along with averaged ART and AutoART thresholds and slopes over all 12 electrodes. A parallel analysis was performed for HS, derived from the difference between baseline and 6-month SI. Finally, correlations between CN CSA and SI, as well as CN CSA and HS were examined. RESULTS: The results of the study showed a significant positive correlation between ART and AutoART ECAP thresholds and as well as slopes for E2, E6, E10 and averaged thresholds and slopes of E1-E12. However, no significant correlation was observed between ART and AutoART averaged ECAP thresholds and slopes and either SI and HS or CN CSA. Furthermore, no significant correlation was found between CN CSA and SI and HS. CONCLUSION: While AutoART is a reliable and safe program for measuring ECAPs in adults, the study found no preoperative prognostic information on intraoperative ECAP results using parameters extracted from current MRI sequences or pre-/intraoperative information on subsequent hearing outcome using ECAP and CN CSA.


Assuntos
Implantes Cocleares , Nervo Coclear , Potenciais Evocados Auditivos , Imageamento por Ressonância Magnética , Humanos , Nervo Coclear/diagnóstico por imagem , Estudos Retrospectivos , Masculino , Pessoa de Meia-Idade , Feminino , Adulto , Idoso , Imageamento por Ressonância Magnética/métodos , Potenciais Evocados Auditivos/fisiologia , Implante Coclear/métodos , Telemetria/métodos , Inteligibilidade da Fala/fisiologia , Adulto Jovem , Valor Preditivo dos Testes , Limiar Auditivo/fisiologia , Potenciais de Ação/fisiologia
10.
Cortex ; 172: 54-71, 2024 03.
Artigo em Inglês | MEDLINE | ID: mdl-38215511

RESUMO

Cortical tracking of speech is vital for speech segmentation and is linked to speech intelligibility. However, there is no clear consensus as to whether reduced intelligibility leads to a decrease or an increase in cortical speech tracking, warranting further investigation of the factors influencing this relationship. One such factor is listening effort, defined as the cognitive resources necessary for speech comprehension, and reported to have a strong negative correlation with speech intelligibility. Yet, no studies have examined the relationship between speech intelligibility, listening effort, and cortical tracking of speech. The aim of the present study was thus to examine these factors in quiet and distinct adverse listening conditions. Forty-nine normal hearing adults listened to sentences produced casually, presented in quiet and two adverse listening conditions: cafeteria noise and reverberant speech. Electrophysiological responses were registered with electroencephalogram, and listening effort was estimated subjectively using self-reported scores and objectively using pupillometry. Results indicated varying impacts of adverse conditions on intelligibility, listening effort, and cortical tracking of speech, depending on the preservation of the speech temporal envelope. The more distorted envelope in the reverberant condition led to higher listening effort, as reflected in higher subjective scores, increased pupil diameter, and stronger cortical tracking of speech in the delta band. These findings suggest that using measures of listening effort in addition to those of intelligibility is useful for interpreting cortical tracking of speech results. Moreover, reading and phonological skills of participants were positively correlated with listening effort in the cafeteria condition, suggesting a special role of expert language skills in processing speech in this noisy condition. Implications for future research and theories linking atypical cortical tracking of speech and reading disorders are further discussed.


Assuntos
Esforço de Escuta , Percepção da Fala , Adulto , Humanos , Ruído , Cognição/fisiologia , Compreensão , Inteligibilidade da Fala/fisiologia , Percepção da Fala/fisiologia
11.
Cerebellum ; 23(4): 1490-1497, 2024 Aug.
Artigo em Inglês | MEDLINE | ID: mdl-38285133

RESUMO

Dysarthria is disabling in persons with degenerative ataxia. There is limited evidence for speech therapy interventions. In this pilot study, we used the Voice trainer app, which was originally developed for patients with Parkinson's disease, as a feedback tool for vocal control. We hypothesized that patients with ataxic dysarthria would benefit from the Voice trainer app to better control their loudness and pitch, resulting in a lower speaking rate and better intelligibility. This intervention study consisted of five therapy sessions of 30 min within 3 weeks using the principles of the Pitch Limiting Voice Treatment. Patients received real-time visual feedback on loudness and pitch during the exercises. Besides, they were encouraged to practice at home or to use the Voice trainer in daily life. We used observer-rated and patient-rated outcome measures. The primary outcome measure was intelligibility, as measured by the Dutch sentence intelligibility test. Twenty-one out of 25 included patients with degenerative ataxia completed the therapy. We found no statistically significant improvements in intelligibility (p = .56). However, after the intervention, patients were speaking slower (p = .03) and the pause durations were longer (p < .001). The patients were satisfied about using the app. At the group level, we found no evidence for an effect of the Voice trainer app on intelligibility in degenerative ataxia. Because of the heterogeneity of ataxic dysarthria, a more tailor-made rather than generic intervention seems warranted.


Assuntos
Disartria , Aplicativos Móveis , Treinamento da Voz , Humanos , Projetos Piloto , Masculino , Feminino , Pessoa de Meia-Idade , Idoso , Disartria/terapia , Disartria/reabilitação , Adulto , Fonoterapia/métodos , Inteligibilidade da Fala/fisiologia , Resultado do Tratamento
12.
Ear Hear ; 45(2): 425-440, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-37882091

RESUMO

OBJECTIVES: The listening demand incurred by speech perception fluctuates in normal conversation. At the acoustic-phonetic level, natural variation in pronunciation acts as speedbumps to accurate lexical selection. Any given utterance may be more or less phonetically ambiguous-a problem that must be resolved by the listener to choose the correct word. This becomes especially apparent when considering two common speech registers-clear and casual-that have characteristically different levels of phonetic ambiguity. Clear speech prioritizes intelligibility through hyperarticulation which results in less ambiguity at the phonetic level, while casual speech tends to have a more collapsed acoustic space. We hypothesized that listeners would invest greater cognitive resources while listening to casual speech to resolve the increased amount of phonetic ambiguity, as compared with clear speech. To this end, we used pupillometry as an online measure of listening effort during perception of clear and casual continuous speech in two background conditions: quiet and noise. DESIGN: Forty-eight participants performed a probe detection task while listening to spoken, nonsensical sentences (masked and unmasked) while recording pupil size. Pupil size was modeled using growth curve analysis to capture the dynamics of the pupil response as the sentence unfolded. RESULTS: Pupil size during listening was sensitive to the presence of noise and speech register (clear/casual). Unsurprisingly, listeners had overall larger pupil dilations during speech perception in noise, replicating earlier work. The pupil dilation pattern for clear and casual sentences was considerably more complex. Pupil dilation during clear speech trials was slightly larger than for casual speech, across quiet and noisy backgrounds. CONCLUSIONS: We suggest that listener motivation could explain the larger pupil dilations to clearly spoken speech. We propose that, bounded by the context of this task, listeners devoted more resources to perceiving the speech signal with the greatest acoustic/phonetic fidelity. Further, we unexpectedly found systematic differences in pupil dilation preceding the onset of the spoken sentences. Together, these data demonstrate that the pupillary system is not merely reactive but also adaptive-sensitive to both task structure and listener motivation to maximize accurate perception in a limited resource system.


Assuntos
Pupila , Percepção da Fala , Humanos , Pupila/fisiologia , Fala , Ruído , Cognição , Percepção da Fala/fisiologia , Inteligibilidade da Fala/fisiologia
13.
Clin Linguist Phon ; 38(3): 227-248, 2024 03 03.
Artigo em Inglês | MEDLINE | ID: mdl-37122073

RESUMO

The purpose of this study was to examine how neurodegeneration secondary to amyotrophic lateral sclerosis (ALS) impacts speech sound accuracy over time and how speech sound accuracy, in turn, is related to speech intelligibility. Twenty-one participants with ALS read the Bamboo Passage over multiple data collection sessions across several months. Phonemic and orthographic transcriptions were completed for all speech samples. The percentage of phonemes accurately produced was calculated across each phoneme, sound class (i.e. consonants versus vowels), and distinctive feature (i.e. features involved in Manner of Articulation, Place of Articulation, Laryngeal Voicing, Tongue Height, and Tongue Advancement). Intelligibility was determined by calculating the percentage of words correctly transcribed orthographically by naive listeners. Linear mixed effects models were conducted to assess the decline of each distinctive feature over time and its impact on intelligibility. The results demonstrated that overall phonemic production accuracy had a nonlinear relationship with speech intelligibility and that a subset of features (i.e. those dependent on precise lingual and labial constriction and/or extensive lingual and labial movement) were more important for intelligibility and were more impacted over time than other features. Furthermore, findings revealed that consonants were more strongly associated with intelligibility than vowels, but consonants did not significantly differ from vowels in their decline over time. These findings have the potential to (1) strengthen mechanistic understanding of the physiological constraints imposed by neuronal degeneration on speech production and (2) inform the timing and selection of treatment and assessment targets for individuals with ALS.


Assuntos
Esclerose Lateral Amiotrófica , Voz , Humanos , Inteligibilidade da Fala/fisiologia , Fonética , Esclerose Lateral Amiotrófica/complicações , Movimento , Medida da Produção da Fala
14.
J Acoust Soc Am ; 154(6): 3973-3985, 2023 12 01.
Artigo em Inglês | MEDLINE | ID: mdl-38149818

RESUMO

Face masks offer essential protection but also interfere with speech communication. Here, audio-only sentences spoken through four types of masks were presented in noise to young adult listeners. Pupil dilation (an index of cognitive demand), intelligibility, and subjective effort and performance ratings were collected. Dilation increased in response to each mask relative to the no-mask condition and differed significantly where acoustic attenuation was most prominent. These results suggest that the acoustic impact of the mask drives not only the intelligibility of speech, but also the cognitive demands of listening. Subjective effort ratings reflected the same trends as the pupil data.


Assuntos
Máscaras , Percepção da Fala , Adulto Jovem , Humanos , Inteligibilidade da Fala/fisiologia , Ruído/efeitos adversos , Pupila/fisiologia , Cognição , Percepção da Fala/fisiologia
15.
Artigo em Inglês | MEDLINE | ID: mdl-37938964

RESUMO

Dysarthria, a speech disorder often caused by neurological damage, compromises the control of vocal muscles in patients, making their speech unclear and communication troublesome. Recently, voice-driven methods have been proposed to improve the speech intelligibility of patients with dysarthria. However, most methods require a significant representation of both the patient's and target speaker's corpus, which is problematic. This study aims to propose a data augmentation-based voice conversion (VC) system to reduce the recording burden on the speaker. We propose dysarthria voice conversion 3.1 (DVC 3.1) based on a data augmentation approach, including text-to-speech and StarGAN-VC architecture, to synthesize a large target and patient-like corpus to lower the burden of recording. An objective evaluation metric of the Google automatic speech recognition (Google ASR) system and a listening test were used to demonstrate the speech intelligibility benefits of DVC 3.1 under free-talk conditions. The DVC system without data augmentation (DVC 3.0) was used for comparison. Subjective and objective evaluation based on the experimental results indicated that the proposed DVC 3.1 system enhanced the Google ASR of two dysarthria patients by approximately [62.4%, 43.3%] and [55.9%, 57.3%] compared to unprocessed dysarthria speech and the DVC 3.0 system, respectively. Further, the proposed DVC 3.1 increased the speech intelligibility of two dysarthria patients by approximately [54.2%, 22.3%] and [63.4%, 70.1%] compared to unprocessed dysarthria speech and the DVC 3.0 system, respectively. The proposed DVC 3.1 system offers significant potential to improve the speech intelligibility performance of patients with dysarthria and enhance verbal communication quality.


Assuntos
Disartria , Voz , Humanos , Disartria/etiologia , Inteligibilidade da Fala/fisiologia , Músculos Laríngeos
16.
Trends Hear ; 27: 23312165231205107, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-37941413

RESUMO

The current review examines listening effort to uncover how it is implicated in bilingual performance under adverse listening conditions. Various measures of listening effort, including physiological, behavioral, and subjective measures, have been employed to examine listening effort in bilingual children and adults. Adverse listening conditions, stemming from environmental factors, as well as factors related to the speaker or listener, have been examined. The existing literature, although relatively limited to date, points to increased listening effort among bilinguals in their nondominant second language (L2) compared to their dominant first language (L1) and relative to monolinguals. Interestingly, increased effort is often observed even when speech intelligibility remains unaffected. These findings emphasize the importance of considering listening effort alongside speech intelligibility. Building upon the insights gained from the current review, we propose that various factors may modulate the observed effects. These include the particular measure selected to examine listening effort, the characteristics of the adverse condition, as well as factors related to the particular linguistic background of the bilingual speaker. Critically, further research is needed to better understand the impact of these factors on listening effort. The review outlines avenues for future research that would promote a comprehensive understanding of listening effort in bilingual individuals.


Assuntos
Esforço de Escuta , Percepção da Fala , Adulto , Criança , Humanos , Inteligibilidade da Fala/fisiologia , Percepção da Fala/fisiologia
17.
Proc Natl Acad Sci U S A ; 120(49): e2309166120, 2023 Dec 05.
Artigo em Inglês | MEDLINE | ID: mdl-38032934

RESUMO

Neural speech tracking has advanced our understanding of how our brains rapidly map an acoustic speech signal onto linguistic representations and ultimately meaning. It remains unclear, however, how speech intelligibility is related to the corresponding neural responses. Many studies addressing this question vary the level of intelligibility by manipulating the acoustic waveform, but this makes it difficult to cleanly disentangle the effects of intelligibility from underlying acoustical confounds. Here, using magnetoencephalography recordings, we study neural measures of speech intelligibility by manipulating intelligibility while keeping the acoustics strictly unchanged. Acoustically identical degraded speech stimuli (three-band noise-vocoded, ~20 s duration) are presented twice, but the second presentation is preceded by the original (nondegraded) version of the speech. This intermediate priming, which generates a "pop-out" percept, substantially improves the intelligibility of the second degraded speech passage. We investigate how intelligibility and acoustical structure affect acoustic and linguistic neural representations using multivariate temporal response functions (mTRFs). As expected, behavioral results confirm that perceived speech clarity is improved by priming. mTRFs analysis reveals that auditory (speech envelope and envelope onset) neural representations are not affected by priming but only by the acoustics of the stimuli (bottom-up driven). Critically, our findings suggest that segmentation of sounds into words emerges with better speech intelligibility, and most strongly at the later (~400 ms latency) word processing stage, in prefrontal cortex, in line with engagement of top-down mechanisms associated with priming. Taken together, our results show that word representations may provide some objective measures of speech comprehension.


Assuntos
Inteligibilidade da Fala , Percepção da Fala , Inteligibilidade da Fala/fisiologia , Estimulação Acústica/métodos , Fala/fisiologia , Ruído , Acústica , Magnetoencefalografia/métodos , Percepção da Fala/fisiologia
18.
J Speech Lang Hear Res ; 66(10): 4083-4099, 2023 10 04.
Artigo em Inglês | MEDLINE | ID: mdl-37699194

RESUMO

PURPOSE: Current evidence regarding the influence of hearing loss on the pupil response elicited by speech perception is inconsistent. This might be partially due to confounding effects of age. This study aimed to compare pupil responses in age-matched groups of normal-hearing (NH) and hard of hearing (HH) listeners during listening to speech. METHOD: We tested the baseline pupil size and mean and peak pupil dilation response of 17 NH participants (Mage = 46 years; age range: 20-62 years) and 17 HH participants (Mage = 45 years; age range: 20-63 years) who were pairwise matched on age and educational level. Participants performed three speech perception tasks at a 50% intelligibility level: noise-vocoded speech and speech masked with either stationary noise or interfering speech. They also listened to speech presented in quiet. RESULTS: Hearing loss was associated with poorer speech perception, except for noise-vocoded speech. In contrast to NH participants, performance of HH participants did not improve across trials for the interfering speech condition, and it decreased for speech in stationary noise. HH participants had a smaller mean pupil dilation in degraded speech conditions compared to NH participants, but not for speech in quiet. They also had a steeper decline in the baseline pupil size across trials. The baseline pupil size was smaller for noise-vocoded speech as compared to the other conditions. The normalized data showed an additional group effect on the baseline pupil response. CONCLUSIONS: Hearing loss is associated with a smaller pupil response and steeper decline in baseline pupil size during the perception of degraded speech. This suggests difficulties of the HH participants to sustain their effort investment and performance across the test session.


Assuntos
Surdez , Perda Auditiva , Percepção da Fala , Humanos , Adulto Jovem , Adulto , Pessoa de Meia-Idade , Pupila/fisiologia , Inteligibilidade da Fala/fisiologia , Ruído , Percepção da Fala/fisiologia
19.
Sci Rep ; 13(1): 11462, 2023 07 15.
Artigo em Inglês | MEDLINE | ID: mdl-37454168

RESUMO

Bilateral subthalamic nucleus deep brain stimulation (STN-DBS) is an effective treatment in advanced Parkinson's Disease (PD). However, the effects of STN-DBS on speech are still debated, particularly in the long-term follow-up. The objective of this study was to evaluate the long-term effects of bilateral STN-DBS on speech in a cohort of advanced PD patients treated with bilateral STN-DBS. Each patient was assessed before surgery through a neurological evaluation and a perceptual-acoustic analysis of speech and re-assessed in the long-term in different stimulation and drug conditions. The primary outcome was the percentage change of speech intelligibility obtained by comparing the postoperative on-stimulation/off-medication condition with the preoperative off-medication condition. Twenty-five PD patients treated with bilateral STN-DBS with a 5-year follow-up were included. In the long-term, speech intelligibility stayed at the same level as preoperative values when compared with preoperative values. STN-DBS induced a significant acute improvement of speech intelligibility (p < 0.005) in the postoperative assessment when compared to the on-stimulation/off-medication and off-stimulation/off-medication conditions. These results highlight that STN-DBS may handle speech intelligibility even in the long-term.


Assuntos
Estimulação Encefálica Profunda , Doença de Parkinson , Núcleo Subtalâmico , Humanos , Doença de Parkinson/cirurgia , Estimulação Encefálica Profunda/métodos , Resultado do Tratamento , Inteligibilidade da Fala/fisiologia
20.
J Speech Lang Hear Res ; 66(8): 2622-2642, 2023 08 03.
Artigo em Inglês | MEDLINE | ID: mdl-37486782

RESUMO

PURPOSE: In current clinical practice, intelligibility of dysarthric speech is commonly assessed by speech-language therapists (SLTs), in most cases by the therapist caring for the patient being diagnosed. Since SLTs are familiar with dysarthria in general and with the speech of the individual patient to be assessed in particular, they have an adaptation advantage in understanding the patient's utterances. We examined whether and how listeners' assessments of communication-related speech parameters vary as a function of their familiarity with dysarthria in general and with the diagnosed patients in particular. METHOD: Intelligibility, speech naturalness, and perceived listener effort were assessed in 20 persons with dysarthria (PWD). Patients' speech samples were judged by the individual treating therapists, five dysarthria experts who were unfamiliar with the patients, and crowdsourced naïve listeners. Adaptation effects were analyzed using (a) linear mixed models of overall scoring levels, (b) regression models of severity dependence, (c) network analyses of between-listener and between-parameter relationships, and (d) measures of intra- and interobserver consistency. RESULTS: Significant advantages of dysarthria experts over laypeople were found in all parameters. An overall advantage of the treating therapists over nonfamiliar experts was only seen in listening effort. Severity-dependent adaptation effects occurred in all parameters. The therapists' responses were heterogeneous and inconsistent with those of the unfamiliar experts and the naïve listeners. CONCLUSIONS: The way SLTs evaluate communication-relevant speech parameters of the PWD whom they care for is influenced not only by adaptation benefits but also by therapeutic biases. This finding weakens the validity of assessments of communication-relevant speech parameters by the treating therapists themselves and encourages the development and use of alternative methods.


Assuntos
Disartria , Percepção da Fala , Humanos , Disartria/etiologia , Disartria/terapia , Disartria/diagnóstico , Inteligibilidade da Fala/fisiologia , Adaptação Fisiológica , Cognição , Medida da Produção da Fala
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA