Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 832
Filter
1.
Codas ; 36(5): e20230194, 2024.
Article in English | MEDLINE | ID: mdl-39230179

ABSTRACT

PURPOSE: To describe the effects of subthalamic nucleus deep brain stimulation (STN-DBS) on the speech of Spanish-speaking Parkinson's disease (PD) patients during the first year of treatment. METHODS: The speech measures (SMs): maximum phonation time, acoustic voice measures, speech rate, speech intelligibility measures, and oral diadochokinesis rates of nine Colombian idiopathic PD patients (four females and five males; age = 63 ± 7 years; years of PD = 10 ± 7 years; UPDRS-III = 57 ± 6; H&Y = 2 ± 0.3) were studied in OFF and ON medication states before and every three months during the first year after STN-DBS surgery. Praat software and healthy native listeners' ratings were used for speech analysis. Statistical analysis tried to find significant differences in the SMs during follow-up (Friedman test) and between medication states (Wilcoxon paired test). Also, a pre-surgery variation interval (PSVI) of reference for every participant and SM was calculated to make an individual analysis of post-surgery variation. RESULTS: Non-significative post-surgery or medication state-related differences in the SMs were found. Nevertheless, individually, based on PSVIs, the SMs exhibited: no variation, inconsistent or consistent variation during post-surgery follow-up in different combinations, depending on the medication state. CONCLUSION: As a group, participants did not have a shared post-surgery pattern of change in any SM. Instead, based on PSVIs, the SMs varied differently in every participant, which suggests that in Spanish-speaking PD patients, the effects of STN-DBS on speech during the first year of treatment could be highly variable.


Subject(s)
Deep Brain Stimulation , Parkinson Disease , Subthalamic Nucleus , Humans , Parkinson Disease/therapy , Parkinson Disease/physiopathology , Male , Female , Middle Aged , Aged , Speech Intelligibility/physiology , Language , Speech Disorders/etiology , Speech Disorders/therapy , Speech/physiology , Speech Production Measurement , Treatment Outcome
2.
PLoS One ; 19(9): e0307158, 2024.
Article in English | MEDLINE | ID: mdl-39292701

ABSTRACT

This study aimed to investigate integration of alternating speech, a stimulus which classically produces a V-shaped speech intelligibility function with minimum at 2-6 Hz in typical-hearing (TH) listeners. We further studied how degraded speech impacts intelligibility across alternating rates (2, 4, 8, and 32 Hz) using vocoded speech, either in the right ear or bilaterally, to simulate single-sided deafness with a cochlear implant (SSD-CI) and bilateral CIs (BiCI), respectively. To assess potential cortical signatures of across-ear integration, we recorded activity in the bilateral auditory cortices (AC) and dorsolateral prefrontal cortices (DLPFC) during the task using functional near-infrared spectroscopy (fNIRS). For speech intelligibility, the V-shaped function was reproduced only in the BiCI condition; TH (with ceiling scores) and SSD-CI conditions had significantly higher scores across all alternating rates compared to the BiCI condition. For fNIRS, the AC and DLPFC exhibited significantly different activity across alternating rates in the TH condition, with altered activity patterns in both regions in the SSD-CI and BiCI conditions. Our results suggest that degraded speech inputs in one or both ears impact across-ear integration and that different listening strategies were employed for speech integration manifested as differences in cortical activity across conditions.


Subject(s)
Auditory Cortex , Cochlear Implants , Spectroscopy, Near-Infrared , Speech Perception , Humans , Spectroscopy, Near-Infrared/methods , Male , Female , Adult , Speech Perception/physiology , Auditory Cortex/physiology , Auditory Cortex/diagnostic imaging , Young Adult , Speech Intelligibility/physiology , Acoustic Stimulation , Dorsolateral Prefrontal Cortex/physiology , Deafness/physiopathology , Speech/physiology
3.
J Speech Lang Hear Res ; 67(9): 2951-2963, 2024 Sep 12.
Article in English | MEDLINE | ID: mdl-39116309

ABSTRACT

PURPOSE: Parkinson's disease (PD) results in hypokinetic dysarthria in as many as 90% of cases. Among the most distinctive features of hypokinetic dysarthria are atypical timing and articulatory imprecision in speech production. Here, we examined the contributions of perceived speech timing typicality and articulatory precision, both on their own and while controlling for the other, on intelligibility and naturalness in speakers with PD. METHOD: Twenty speakers with PD and four healthy older adults read aloud the first paragraph of the Rainbow Passage. Twenty inexperienced listeners with typical hearing listened to these recordings and rated intelligibility, naturalness, timing typicality, and articulatory precision using separate visual analog scales. Ratings were averaged across listeners and entered into linear regression models with intelligibility and naturalness as dependent variables and timing typicality and articulatory precision as independent variables in each. RESULTS: Articulatory precision, but not timing typicality, was positively correlated with intelligibility on its own, but neither was associated with intelligibility after accounting for the other. Both timing typicality and articulatory precision were positively correlated with naturalness on their own as well as after controlling for the other variable. CONCLUSION: These results contribute to the overall understanding of speech factors associated with intelligibility and naturalness in speakers with PD and indicate that considering the unique contributions of related perceptual constructs may provide more information than bivariate relationships alone.


Subject(s)
Dysarthria , Parkinson Disease , Speech Intelligibility , Speech Perception , Humans , Parkinson Disease/physiopathology , Parkinson Disease/complications , Parkinson Disease/psychology , Speech Intelligibility/physiology , Male , Female , Aged , Speech Perception/physiology , Middle Aged , Dysarthria/etiology , Dysarthria/physiopathology , Time Factors , Speech Production Measurement , Aged, 80 and over , Speech/physiology
4.
Trends Hear ; 28: 23312165241273346, 2024.
Article in English | MEDLINE | ID: mdl-39195628

ABSTRACT

There is broad consensus that listening effort is an important outcome for measuring hearing performance. However, there remains debate on the best ways to measure listening effort. This study sought to measure neural correlates of listening effort using functional near-infrared spectroscopy (fNIRS) in experienced adult hearing aid users. The study evaluated impacts of amplification and signal-to-noise ratio (SNR) on cerebral blood oxygenation, with the expectation that easier listening conditions would be associated with less oxygenation in the prefrontal cortex. Thirty experienced adult hearing aid users repeated sentence-final words from low-context Revised Speech Perception in Noise Test sentences. Participants repeated words at a hard SNR (individual SNR-50) or easy SNR (individual SNR-50 + 10 dB), while wearing hearing aids fit to prescriptive targets or without wearing hearing aids. In addition to assessing listening accuracy and subjective listening effort, prefrontal blood oxygenation was measured using fNIRS. As expected, easier listening conditions (i.e., easy SNR, with hearing aids) led to better listening accuracy, lower subjective listening effort, and lower oxygenation across the entire prefrontal cortex compared to harder listening conditions. Listening accuracy and subjective listening effort were also significant predictors of oxygenation.


Subject(s)
Hearing Aids , Spectroscopy, Near-Infrared , Speech Perception , Humans , Male , Female , Speech Perception/physiology , Aged , Middle Aged , Signal-To-Noise Ratio , Acoustic Stimulation/methods , Prefrontal Cortex/physiology , Persons With Hearing Impairments/psychology , Persons With Hearing Impairments/rehabilitation , Noise/adverse effects , Correction of Hearing Impairment/instrumentation , Correction of Hearing Impairment/methods , Adult , Aged, 80 and over , Hearing/physiology , Cerebrovascular Circulation/physiology , Auditory Threshold/physiology , Speech Intelligibility/physiology
5.
Sci Rep ; 14(1): 20069, 2024 08 29.
Article in English | MEDLINE | ID: mdl-39209957

ABSTRACT

Communication is a fundamental aspect of human interaction, yet many individuals must speak in less-than-ideal acoustic environments daily. Adapting their speech to ensure intelligibility in these varied settings can impose a significant cognitive burden. Understanding this burden on talkers has significant implications for the design of public spaces and workplace environments, as well as speaker training programs. The aim of this study was to examine how room acoustics and speaking style affect cognitive load through self-rating of mental demand and pupillometry. Nineteen adult native speakers of American English were instructed to read sentences in both casual and clear speech-a technique known to enhance intelligibility-across three levels of reverberation (0.05 s, 1.2 s, and 1.83 s at 500-1000 Hz). Our findings revealed that speaking style consistently affects the cognitive load on talkers more than room acoustics across the tested reverberation range. Specifically, pupillometry data suggested that speaking in clear speech elevates the cognitive load comparably to speaking in a room with long reverberation, challenging the conventional view of clear speech as an 'easy' strategy for improving intelligibility. These results underscore the importance of accounting for talkers' cognitive load when optimizing room acoustics and developing speech production training.


Subject(s)
Cognition , Speech Intelligibility , Humans , Male , Female , Cognition/physiology , Adult , Speech Intelligibility/physiology , Speech/physiology , Young Adult , Speech Perception/physiology , Acoustics
6.
J Int Adv Otol ; 20(4): 289-300, 2024 Jul 29.
Article in English | MEDLINE | ID: mdl-39159037

ABSTRACT

People with single-sided deafness (SSD) or asymmetric hearing loss (AHL) have particular difficulty understanding speech in noisy listening situations and in sound localization. The objective of this multicenter study is to evaluate the effect of a cochlear implant (CI) in adults with single-sided deafness (SSD) or asymmetric hearing loss (AHL), particularly regarding sound localization and speech intelligibility with additional interest in electric-acoustic pitch matching. A prospective longitudinal study at 7 European tertiary referral centers was conducted including 19 SSD and 16 AHL subjects undergoing cochlear implantation. Sound localization accuracy was investigated in terms of root mean square error and signed bias before and after implantation. Speech recognition in quiet and speech reception thresholds in noise for several spatial configurations were assessed preoperatively and at several post-activation time points. Pitch perception with CI was tracked using pitch matching. Data up to 12 months post activation were collected. In both SSD and AHL subjects, CI significantly improved sound localization for sound sources on the implant side, and thus overall sound localization. Speech recognition in quiet with the implant ear improved significantly. In noise, a significant head shadow effect was found for SSD subjects only. However, the evaluation of AHL subjects was limited by the small sample size. No uniform development of pitch perception with the implant ear was observed. The benefits shown in this study confirm and expand the existing body of evidence for the effectiveness of CI in SSD and AHL. Particularly, improved localization was shown to result from increased localization accuracy on the implant side.


Subject(s)
Cochlear Implantation , Cochlear Implants , Hearing Loss, Unilateral , Sound Localization , Speech Perception , Humans , Cochlear Implantation/methods , Male , Sound Localization/physiology , Female , Middle Aged , Speech Perception/physiology , Prospective Studies , Hearing Loss, Unilateral/surgery , Hearing Loss, Unilateral/rehabilitation , Hearing Loss, Unilateral/physiopathology , Follow-Up Studies , Aged , Adult , Europe , Longitudinal Studies , Treatment Outcome , Speech Intelligibility/physiology , Pitch Perception/physiology , Deafness/surgery , Deafness/rehabilitation , Deafness/physiopathology , Noise
7.
Proc Natl Acad Sci U S A ; 121(34): e2411167121, 2024 Aug 20.
Article in English | MEDLINE | ID: mdl-39136991

ABSTRACT

Evidence accumulates that the cerebellum's role in the brain is not restricted to motor functions. Rather, cerebellar activity seems to be crucial for a variety of tasks that rely on precise event timing and prediction. Due to its complex structure and importance in communication, human speech requires a particularly precise and predictive coordination of neural processes to be successfully comprehended. Recent studies proposed that the cerebellum is indeed a major contributor to speech processing, but how this contribution is achieved mechanistically remains poorly understood. The current study aimed to reveal a mechanism underlying cortico-cerebellar coordination and demonstrate its speech-specificity. In a reanalysis of magnetoencephalography data, we found that activity in the cerebellum aligned to rhythmic sequences of noise-vocoded speech, irrespective of its intelligibility. We then tested whether these "entrained" responses persist, and how they interact with other brain regions, when a rhythmic stimulus stopped and temporal predictions had to be updated. We found that only intelligible speech produced sustained rhythmic responses in the cerebellum. During this "entrainment echo," but not during rhythmic speech itself, cerebellar activity was coupled with that in the left inferior frontal gyrus, and specifically at rates corresponding to the preceding stimulus rhythm. This finding represents evidence for specific cerebellum-driven temporal predictions in speech processing and their relay to cortical regions.


Subject(s)
Cerebellum , Magnetoencephalography , Humans , Cerebellum/physiology , Male , Female , Adult , Speech Perception/physiology , Young Adult , Speech/physiology , Speech Intelligibility/physiology
8.
Behav Res Methods ; 56(7): 8132-8154, 2024 10.
Article in English | MEDLINE | ID: mdl-39048860

ABSTRACT

When investigating unobservable, complex traits, data collection and aggregation processes can introduce distinctive features to the data such as boundedness, measurement error, clustering, outliers, and heteroscedasticity. Failure to collectively address these features can result in statistical challenges that prevent the investigation of hypotheses regarding these traits. This study aimed to demonstrate the efficacy of the Bayesian beta-proportion generalized linear latent and mixed model (beta-proportion GLLAMM) (Rabe-Hesketh et al., Psychometrika, 69(2), 167-90, 2004a, Journal of Econometrics, 128(2), 301-23, 2004c, 2004b; Skrondal and Rabe-Hesketh 2004) in handling data features when exploring research hypotheses concerning speech intelligibility. To achieve this objective, the study reexamined data from transcriptions of spontaneous speech samples initially collected by Boonen et al. (Journal of Child Language, 50(1), 78-103, 2023). The data were aggregated into entropy scores. The research compared the prediction accuracy of the beta-proportion GLLAMM with the normal linear mixed model (LMM) (Holmes et al., 2019) and investigated its capacity to estimate a latent intelligibility from entropy scores. The study also illustrated how hypotheses concerning the impact of speaker-related factors on intelligibility can be explored with the proposed model. The beta-proportion GLLAMM was not free of challenges; its implementation required formulating assumptions about the data-generating process and knowledge of probabilistic programming languages, both central to Bayesian methods. Nevertheless, results indicated the superiority of the model in predicting empirical phenomena over the normal LMM, and its ability to quantify a latent potential intelligibility. Additionally, the proposed model facilitated the exploration of hypotheses concerning speaker-related factors and intelligibility. Ultimately, this research has implications for researchers and data analysts interested in quantitatively measuring intricate, unobservable constructs while accurately predicting the empirical phenomena.


Subject(s)
Bayes Theorem , Entropy , Speech Intelligibility , Humans , Speech Intelligibility/physiology , Linear Models , Models, Statistical , Data Interpretation, Statistical
9.
Trends Hear ; 28: 23312165241240572, 2024.
Article in English | MEDLINE | ID: mdl-38676325

ABSTRACT

Realistic outcome measures that reflect everyday hearing challenges are needed to assess hearing aid and cochlear implant (CI) fitting. Literature suggests that listening effort measures may be more sensitive to differences between hearing-device settings than established speech intelligibility measures when speech intelligibility is near maximum. Which method provides the most effective measurement of listening effort for this purpose is currently unclear. This study aimed to investigate the feasibility of two tests for measuring changes in listening effort in CI users due to signal-to-noise ratio (SNR) differences, as would arise from different hearing-device settings. By comparing the effect size of SNR differences on listening effort measures with test-retest differences, the study evaluated the suitability of these tests for clinical use. Nineteen CI users underwent two listening effort tests at two SNRs (+4 and +8 dB relative to individuals' 50% speech perception threshold). We employed dual-task paradigms-a sentence-final word identification and recall test (SWIRT) and a sentence verification test (SVT)-to assess listening effort at these two SNRs. Our results show a significant difference in listening effort between the SNRs for both test methods, although the effect size was comparable to the test-retest difference, and the sensitivity was not superior to speech intelligibility measures. Thus, the implementations of SVT and SWIRT used in this study are not suitable for clinical use to measure listening effort differences of this magnitude in individual CI users. However, they can be used in research involving CI users to analyze group data.


Subject(s)
Cochlear Implantation , Cochlear Implants , Feasibility Studies , Persons With Hearing Impairments , Speech Intelligibility , Speech Perception , Humans , Male , Female , Speech Perception/physiology , Middle Aged , Aged , Speech Intelligibility/physiology , Cochlear Implantation/instrumentation , Persons With Hearing Impairments/rehabilitation , Persons With Hearing Impairments/psychology , Reproducibility of Results , Acoustic Stimulation , Signal-To-Noise Ratio , Adult , Aged, 80 and over , Auditory Threshold/physiology , Predictive Value of Tests , Correction of Hearing Impairment/instrumentation , Noise/adverse effects
10.
J Speech Lang Hear Res ; 67(4): 1090-1106, 2024 04 08.
Article in English | MEDLINE | ID: mdl-38498664

ABSTRACT

PURPOSE: This study examined speech changes induced by deep-brain stimulation (DBS) in speakers with Parkinson's disease (PD) using a set of auditory-perceptual and acoustic measures. METHOD: Speech recordings from nine speakers with PD and DBS were compared between DBS-On and DBS-Off conditions using auditory-perceptual and acoustic analyses. Auditory-perceptual ratings included voice quality, articulation precision, prosody, speech intelligibility, and listening effort obtained from 44 listeners. Acoustic measures were made for voicing proportion, second formant frequency slope, vowel dispersion, articulation rate, and range of fundamental frequency and intensity. RESULTS: No significant changes were found between DBS-On and DBS-Off for the five perceptual ratings. Four of six acoustic measures revealed significant differences between the two conditions. While articulation rate and acoustic vowel dispersion increased, voicing proportion and intensity range decreased from the DBS-Off to DBS-On condition. However, a visual examination of the data indicated that the statistical significance was mostly driven by a small number of participants, while the majority did not show a consistent pattern of such changes. CONCLUSIONS: Our data, in general, indicate no-to-minimal changes in speech production ensued from DBS stimulation. The findings are discussed with a focus on large interspeaker variability in PD in terms of their speech characteristics and the potential effects of DBS on speech.


Subject(s)
Deep Brain Stimulation , Parkinson Disease , Humans , Acoustics , Speech Intelligibility/physiology , Voice Quality , Parkinson Disease/complications , Parkinson Disease/therapy , Brain , Speech Acoustics
11.
Trends Hear ; 28: 23312165241232551, 2024.
Article in English | MEDLINE | ID: mdl-38549351

ABSTRACT

In daily life, both acoustic factors and social context can affect listening effort investment. In laboratory settings, information about listening effort has been deduced from pupil and cardiovascular responses independently. The extent to which these measures can jointly predict listening-related factors is unknown. Here we combined pupil and cardiovascular features to predict acoustic and contextual aspects of speech perception. Data were collected from 29 adults (mean  =  64.6 years, SD  =  9.2) with hearing loss. Participants performed a speech perception task at two individualized signal-to-noise ratios (corresponding to 50% and 80% of sentences correct) and in two social contexts (the presence and absence of two observers). Seven features were extracted per trial: baseline pupil size, peak pupil dilation, mean pupil dilation, interbeat interval, blood volume pulse amplitude, pre-ejection period and pulse arrival time. These features were used to train k-nearest neighbor classifiers to predict task demand, social context and sentence accuracy. The k-fold cross validation on the group-level data revealed above-chance classification accuracies: task demand, 64.4%; social context, 78.3%; and sentence accuracy, 55.1%. However, classification accuracies diminished when the classifiers were trained and tested on data from different participants. Individually trained classifiers (one per participant) performed better than group-level classifiers: 71.7% (SD  =  10.2) for task demand, 88.0% (SD  =  7.5) for social context, and 60.0% (SD  =  13.1) for sentence accuracy. We demonstrated that classifiers trained on group-level physiological data to predict aspects of speech perception generalized poorly to novel participants. Individually calibrated classifiers hold more promise for future applications.


Subject(s)
Pupil , Speech Perception , Humans , Pupil/physiology , Speech Intelligibility/physiology , Speech Perception/physiology , Middle Aged , Aged
12.
Otol Neurotol ; 45(5): e385-e392, 2024 Jun 01.
Article in English | MEDLINE | ID: mdl-38518764

ABSTRACT

HYPOTHESIS: The behaviorally based programming with loudness balancing (LB) would result in better speech understanding, spectral-temporal resolution, and music perception scores, and there would be a relationship between these scores. BACKGROUND: Loudness imbalances at upper stimulation levels may cause sounds to be perceived as irregular, gravelly, or overly echoed and may negatively affect the listening performance of the cochlear implant (CI) user. LB should be performed after fitting to overcome these problems. METHODS: The study included 26 unilateral Med-EL CI users. Two different CI programs based on the objective electrically evoked stapedial reflex threshold (P1) and the behaviorally program with LB (P2) were recorded for each participant. The Turkish Matrix Sentence Test (TMS) was applied to evaluate speech perception; the Random Gap Detection Test (RGDT) and Spectral-Temporally Modulated Ripple Test (SMRT) were applied to evaluate spectral temporal resolution skills; the Mini Profile of Music Perception Skills (mini-PROMS) and Melodic Contour Identification (MCI) tests were applied to evaluate music perception, and the results were compared. RESULTS: Significantly better scores were obtained with P2 in TMS tests performed in noise and quiet. SMRT scores were significantly correlated with TMS in quiet and noise, and mini-PROMS sound perception results. Although better scores were obtained with P2 in the mini-PROMS total score and MCI, a significant difference was found only for MCI. CONCLUSION: The data from the current study showed that equalization of loudness across CI electrodes leads to better perceptual acuity. It also revealed the relationship between speech perception, spectral-temporal resolution, and music perception.


Subject(s)
Cochlear Implantation , Cochlear Implants , Music , Speech Perception , Humans , Male , Female , Middle Aged , Adult , Speech Perception/physiology , Cochlear Implantation/methods , Speech Intelligibility/physiology , Aged , Auditory Perception/physiology , Loudness Perception/physiology , Young Adult
13.
JASA Express Lett ; 4(2)2024 Feb 01.
Article in English | MEDLINE | ID: mdl-38350077

ABSTRACT

Measuring how well human listeners recognize speech under varying environmental conditions (speech intelligibility) is a challenge for theoretical, technological, and clinical approaches to speech communication. The current gold standard-human transcription-is time- and resource-intensive. Recent advances in automatic speech recognition (ASR) systems raise the possibility of automating intelligibility measurement. This study tested 4 state-of-the-art ASR systems with second language speech-in-noise and found that one, whisper, performed at or above human listener accuracy. However, the content of whisper's responses diverged substantially from human responses, especially at lower signal-to-noise ratios, suggesting both opportunities and limitations for ASR--based speech intelligibility modeling.


Subject(s)
Speech Perception , Humans , Speech Perception/physiology , Noise/adverse effects , Speech Intelligibility/physiology , Speech Recognition Software , Recognition, Psychology
14.
PLoS Biol ; 22(2): e3002498, 2024 Feb.
Article in English | MEDLINE | ID: mdl-38358954

ABSTRACT

Speech recognition crucially relies on slow temporal modulations (<16 Hz) in speech. Recent studies, however, have demonstrated that the long-delay echoes, which are common during online conferencing, can eliminate crucial temporal modulations in speech but do not affect speech intelligibility. Here, we investigated the underlying neural mechanisms. MEG experiments demonstrated that cortical activity can effectively track the temporal modulations eliminated by an echo, which cannot be fully explained by basic neural adaptation mechanisms. Furthermore, cortical responses to echoic speech can be better explained by a model that segregates speech from its echo than by a model that encodes echoic speech as a whole. The speech segregation effect was observed even when attention was diverted but would disappear when segregation cues, i.e., speech fine structure, were removed. These results strongly suggested that, through mechanisms such as stream segregation, the auditory system can build an echo-insensitive representation of speech envelope, which can support reliable speech recognition.


Subject(s)
Auditory Cortex , Speech Perception , Humans , Speech Perception/physiology , Speech Intelligibility/physiology , Brain , Auditory Cortex/physiology , Attention , Acoustic Stimulation
15.
Cortex ; 172: 54-71, 2024 03.
Article in English | MEDLINE | ID: mdl-38215511

ABSTRACT

Cortical tracking of speech is vital for speech segmentation and is linked to speech intelligibility. However, there is no clear consensus as to whether reduced intelligibility leads to a decrease or an increase in cortical speech tracking, warranting further investigation of the factors influencing this relationship. One such factor is listening effort, defined as the cognitive resources necessary for speech comprehension, and reported to have a strong negative correlation with speech intelligibility. Yet, no studies have examined the relationship between speech intelligibility, listening effort, and cortical tracking of speech. The aim of the present study was thus to examine these factors in quiet and distinct adverse listening conditions. Forty-nine normal hearing adults listened to sentences produced casually, presented in quiet and two adverse listening conditions: cafeteria noise and reverberant speech. Electrophysiological responses were registered with electroencephalogram, and listening effort was estimated subjectively using self-reported scores and objectively using pupillometry. Results indicated varying impacts of adverse conditions on intelligibility, listening effort, and cortical tracking of speech, depending on the preservation of the speech temporal envelope. The more distorted envelope in the reverberant condition led to higher listening effort, as reflected in higher subjective scores, increased pupil diameter, and stronger cortical tracking of speech in the delta band. These findings suggest that using measures of listening effort in addition to those of intelligibility is useful for interpreting cortical tracking of speech results. Moreover, reading and phonological skills of participants were positively correlated with listening effort in the cafeteria condition, suggesting a special role of expert language skills in processing speech in this noisy condition. Implications for future research and theories linking atypical cortical tracking of speech and reading disorders are further discussed.


Subject(s)
Listening Effort , Speech Perception , Adult , Humans , Noise , Cognition/physiology , Comprehension , Speech Intelligibility/physiology , Speech Perception/physiology
16.
Eur Arch Otorhinolaryngol ; 281(7): 3461-3473, 2024 Jul.
Article in English | MEDLINE | ID: mdl-38219245

ABSTRACT

PURPOSE: The purpose of this retrospective study is to compare the results of electrically evoked compound action potential (ECAP) measurements using automatic auditory response telemetry (AutoART) with those obtained by ART in adults. The study also aimed to evaluate the predictive value of intraoperative ART and AutoART ECAPs for speech intelligibility (SI) and hearing success (HS), and to determine if cochlear nerve (CN) cross-sectional area (CSA) obtained preoperatively by magnetic resonance imaging (MRI) scans could predict ART and AutoART ECAPs and SI and HS outcome. METHODS: The study analyzed and correlated ART and AutoART ECAP thresholds at electrodes E2, E6, and E10, as well as averaged ECAP thresholds over electrodes E1-E12, using data from 32 implants. Correlations were also examined for ART and AutoART ECAP slopes. In addition, averaged ART and AutoART ECAP thresholds and slopes over all 12 electrodes for each participant were correlated with CN CSA measured from MRI sequences. SI of the monosyllabic Freiburg Speech Test at 65 dB sound pressure level was examined along with averaged ART and AutoART thresholds and slopes over all 12 electrodes. A parallel analysis was performed for HS, derived from the difference between baseline and 6-month SI. Finally, correlations between CN CSA and SI, as well as CN CSA and HS were examined. RESULTS: The results of the study showed a significant positive correlation between ART and AutoART ECAP thresholds and as well as slopes for E2, E6, E10 and averaged thresholds and slopes of E1-E12. However, no significant correlation was observed between ART and AutoART averaged ECAP thresholds and slopes and either SI and HS or CN CSA. Furthermore, no significant correlation was found between CN CSA and SI and HS. CONCLUSION: While AutoART is a reliable and safe program for measuring ECAPs in adults, the study found no preoperative prognostic information on intraoperative ECAP results using parameters extracted from current MRI sequences or pre-/intraoperative information on subsequent hearing outcome using ECAP and CN CSA.


Subject(s)
Cochlear Implants , Cochlear Nerve , Evoked Potentials, Auditory , Magnetic Resonance Imaging , Humans , Cochlear Nerve/diagnostic imaging , Retrospective Studies , Male , Middle Aged , Female , Adult , Aged , Magnetic Resonance Imaging/methods , Evoked Potentials, Auditory/physiology , Cochlear Implantation/methods , Telemetry/methods , Speech Intelligibility/physiology , Young Adult , Predictive Value of Tests , Auditory Threshold/physiology , Action Potentials/physiology
17.
Cerebellum ; 23(4): 1490-1497, 2024 Aug.
Article in English | MEDLINE | ID: mdl-38285133

ABSTRACT

Dysarthria is disabling in persons with degenerative ataxia. There is limited evidence for speech therapy interventions. In this pilot study, we used the Voice trainer app, which was originally developed for patients with Parkinson's disease, as a feedback tool for vocal control. We hypothesized that patients with ataxic dysarthria would benefit from the Voice trainer app to better control their loudness and pitch, resulting in a lower speaking rate and better intelligibility. This intervention study consisted of five therapy sessions of 30 min within 3 weeks using the principles of the Pitch Limiting Voice Treatment. Patients received real-time visual feedback on loudness and pitch during the exercises. Besides, they were encouraged to practice at home or to use the Voice trainer in daily life. We used observer-rated and patient-rated outcome measures. The primary outcome measure was intelligibility, as measured by the Dutch sentence intelligibility test. Twenty-one out of 25 included patients with degenerative ataxia completed the therapy. We found no statistically significant improvements in intelligibility (p = .56). However, after the intervention, patients were speaking slower (p = .03) and the pause durations were longer (p < .001). The patients were satisfied about using the app. At the group level, we found no evidence for an effect of the Voice trainer app on intelligibility in degenerative ataxia. Because of the heterogeneity of ataxic dysarthria, a more tailor-made rather than generic intervention seems warranted.


Subject(s)
Dysarthria , Mobile Applications , Voice Training , Humans , Pilot Projects , Male , Female , Middle Aged , Aged , Dysarthria/therapy , Dysarthria/rehabilitation , Adult , Speech Therapy/methods , Speech Intelligibility/physiology , Treatment Outcome
18.
Ear Hear ; 45(2): 425-440, 2024.
Article in English | MEDLINE | ID: mdl-37882091

ABSTRACT

OBJECTIVES: The listening demand incurred by speech perception fluctuates in normal conversation. At the acoustic-phonetic level, natural variation in pronunciation acts as speedbumps to accurate lexical selection. Any given utterance may be more or less phonetically ambiguous-a problem that must be resolved by the listener to choose the correct word. This becomes especially apparent when considering two common speech registers-clear and casual-that have characteristically different levels of phonetic ambiguity. Clear speech prioritizes intelligibility through hyperarticulation which results in less ambiguity at the phonetic level, while casual speech tends to have a more collapsed acoustic space. We hypothesized that listeners would invest greater cognitive resources while listening to casual speech to resolve the increased amount of phonetic ambiguity, as compared with clear speech. To this end, we used pupillometry as an online measure of listening effort during perception of clear and casual continuous speech in two background conditions: quiet and noise. DESIGN: Forty-eight participants performed a probe detection task while listening to spoken, nonsensical sentences (masked and unmasked) while recording pupil size. Pupil size was modeled using growth curve analysis to capture the dynamics of the pupil response as the sentence unfolded. RESULTS: Pupil size during listening was sensitive to the presence of noise and speech register (clear/casual). Unsurprisingly, listeners had overall larger pupil dilations during speech perception in noise, replicating earlier work. The pupil dilation pattern for clear and casual sentences was considerably more complex. Pupil dilation during clear speech trials was slightly larger than for casual speech, across quiet and noisy backgrounds. CONCLUSIONS: We suggest that listener motivation could explain the larger pupil dilations to clearly spoken speech. We propose that, bounded by the context of this task, listeners devoted more resources to perceiving the speech signal with the greatest acoustic/phonetic fidelity. Further, we unexpectedly found systematic differences in pupil dilation preceding the onset of the spoken sentences. Together, these data demonstrate that the pupillary system is not merely reactive but also adaptive-sensitive to both task structure and listener motivation to maximize accurate perception in a limited resource system.


Subject(s)
Pupil , Speech Perception , Humans , Pupil/physiology , Speech , Noise , Cognition , Speech Perception/physiology , Speech Intelligibility/physiology
19.
Clin Linguist Phon ; 38(3): 227-248, 2024 03 03.
Article in English | MEDLINE | ID: mdl-37122073

ABSTRACT

The purpose of this study was to examine how neurodegeneration secondary to amyotrophic lateral sclerosis (ALS) impacts speech sound accuracy over time and how speech sound accuracy, in turn, is related to speech intelligibility. Twenty-one participants with ALS read the Bamboo Passage over multiple data collection sessions across several months. Phonemic and orthographic transcriptions were completed for all speech samples. The percentage of phonemes accurately produced was calculated across each phoneme, sound class (i.e. consonants versus vowels), and distinctive feature (i.e. features involved in Manner of Articulation, Place of Articulation, Laryngeal Voicing, Tongue Height, and Tongue Advancement). Intelligibility was determined by calculating the percentage of words correctly transcribed orthographically by naive listeners. Linear mixed effects models were conducted to assess the decline of each distinctive feature over time and its impact on intelligibility. The results demonstrated that overall phonemic production accuracy had a nonlinear relationship with speech intelligibility and that a subset of features (i.e. those dependent on precise lingual and labial constriction and/or extensive lingual and labial movement) were more important for intelligibility and were more impacted over time than other features. Furthermore, findings revealed that consonants were more strongly associated with intelligibility than vowels, but consonants did not significantly differ from vowels in their decline over time. These findings have the potential to (1) strengthen mechanistic understanding of the physiological constraints imposed by neuronal degeneration on speech production and (2) inform the timing and selection of treatment and assessment targets for individuals with ALS.


Subject(s)
Amyotrophic Lateral Sclerosis , Voice , Humans , Speech Intelligibility/physiology , Phonetics , Amyotrophic Lateral Sclerosis/complications , Movement , Speech Production Measurement
20.
J Acoust Soc Am ; 154(6): 3973-3985, 2023 12 01.
Article in English | MEDLINE | ID: mdl-38149818

ABSTRACT

Face masks offer essential protection but also interfere with speech communication. Here, audio-only sentences spoken through four types of masks were presented in noise to young adult listeners. Pupil dilation (an index of cognitive demand), intelligibility, and subjective effort and performance ratings were collected. Dilation increased in response to each mask relative to the no-mask condition and differed significantly where acoustic attenuation was most prominent. These results suggest that the acoustic impact of the mask drives not only the intelligibility of speech, but also the cognitive demands of listening. Subjective effort ratings reflected the same trends as the pupil data.


Subject(s)
Masks , Speech Perception , Young Adult , Humans , Speech Intelligibility/physiology , Noise/adverse effects , Pupil/physiology , Cognition , Speech Perception/physiology
SELECTION OF CITATIONS
SEARCH DETAIL