Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 790
Filtrar
Más filtros

Medicinas Complementárias
Intervalo de año de publicación
1.
Proc Natl Acad Sci U S A ; 120(49): e2309166120, 2023 Dec 05.
Artículo en Inglés | MEDLINE | ID: mdl-38032934

RESUMEN

Neural speech tracking has advanced our understanding of how our brains rapidly map an acoustic speech signal onto linguistic representations and ultimately meaning. It remains unclear, however, how speech intelligibility is related to the corresponding neural responses. Many studies addressing this question vary the level of intelligibility by manipulating the acoustic waveform, but this makes it difficult to cleanly disentangle the effects of intelligibility from underlying acoustical confounds. Here, using magnetoencephalography recordings, we study neural measures of speech intelligibility by manipulating intelligibility while keeping the acoustics strictly unchanged. Acoustically identical degraded speech stimuli (three-band noise-vocoded, ~20 s duration) are presented twice, but the second presentation is preceded by the original (nondegraded) version of the speech. This intermediate priming, which generates a "pop-out" percept, substantially improves the intelligibility of the second degraded speech passage. We investigate how intelligibility and acoustical structure affect acoustic and linguistic neural representations using multivariate temporal response functions (mTRFs). As expected, behavioral results confirm that perceived speech clarity is improved by priming. mTRFs analysis reveals that auditory (speech envelope and envelope onset) neural representations are not affected by priming but only by the acoustics of the stimuli (bottom-up driven). Critically, our findings suggest that segmentation of sounds into words emerges with better speech intelligibility, and most strongly at the later (~400 ms latency) word processing stage, in prefrontal cortex, in line with engagement of top-down mechanisms associated with priming. Taken together, our results show that word representations may provide some objective measures of speech comprehension.


Asunto(s)
Inteligibilidad del Habla , Percepción del Habla , Inteligibilidad del Habla/fisiología , Estimulación Acústica/métodos , Habla/fisiología , Ruido , Acústica , Magnetoencefalografía/métodos , Percepción del Habla/fisiología
2.
Neuroimage ; 278: 120271, 2023 09.
Artículo en Inglés | MEDLINE | ID: mdl-37442310

RESUMEN

Humans have the unique ability to decode the rapid stream of language elements that constitute speech, even when it is contaminated by noise. Two reliable observations about noisy speech perception are that seeing the face of the talker improves intelligibility and the existence of individual differences in the ability to perceive noisy speech. We introduce a multivariate BOLD fMRI measure that explains both observations. In two independent fMRI studies, clear and noisy speech was presented in visual, auditory and audiovisual formats to thirty-seven participants who rated intelligibility. An event-related design was used to sort noisy speech trials by their intelligibility. Individual-differences multidimensional scaling was applied to fMRI response patterns in superior temporal cortex and the dissimilarity between responses to clear speech and noisy (but intelligible) speech was measured. Neural dissimilarity was less for audiovisual speech than auditory-only speech, corresponding to the greater intelligibility of noisy audiovisual speech. Dissimilarity was less in participants with better noisy speech perception, corresponding to individual differences. These relationships held for both single word and entire sentence stimuli, suggesting that they were driven by intelligibility rather than the specific stimuli tested. A neural measure of perceptual intelligibility may aid in the development of strategies for helping those with impaired speech perception.


Asunto(s)
Percepción del Habla , Habla , Humanos , Imagen por Resonancia Magnética , Individualidad , Percepción Visual/fisiología , Percepción del Habla/fisiología , Lóbulo Temporal/diagnóstico por imagen , Lóbulo Temporal/fisiología , Inteligibilidad del Habla , Estimulación Acústica/métodos
3.
J Speech Lang Hear Res ; 66(8): 2643-2655, 2023 08 03.
Artículo en Inglés | MEDLINE | ID: mdl-37499233

RESUMEN

PURPOSE: Pitch variations of the fundamental frequency (fo) contour contribute to speech perception in noisy environments, but whether musicians confer an advantage in speech in noise (SIN) with altered fo information remains unclear. This study investigated the effects of different levels of degraded fo contour (i.e., conveying lexical tone or intonation information) on musician advantage in speech-in-noise perception. METHOD: A cohort of native Mandarin Chinese speakers, comprising 30 trained musicians and 30 nonmusicians, were tested on the intelligibility of Mandarin Chinese sentences with natural, flattened-tone, flattened-intonation, and flattened-all fo contours embedded in background noise masked under three signal-to-noise ratios (0, -5, and -9 dB). Pitch difference thresholds and innate musical skills associated with speech-in-noise benefits were also assessed. RESULTS: Speech intelligibility score improved with increasing signal-to-noise level for both musicians and nonmusicians. However, no musician advantage was observed for identifying any type of flattened-fo contour SIN. Musicians exhibited smaller fo pitch discrimination limens than nonmusicians, which correlated with benefits for perceiving speech with intact tone-level fo information. Regardless of musician status, performance on the pitch and accent musical-skill subtests correlated with speech intelligibility score. CONCLUSIONS: Collectively, these results provide no evidence for a musician advantage for perceiving speech with distorted fo information in noisy environments. Results further show that perceptual musical skills on pitch and accent processing may benefit the perception of SIN, independent of formal musical training. Our findings suggest that the potential application of music training in speech perception in noisy backgrounds is not contingent on the ability to process fo pitch contours, at least for Mandarin Chinese speakers. SUPPLEMENTAL MATERIAL: https://doi.org/10.23641/asha.23706354.


Asunto(s)
Música , Percepción del Habla , Humanos , Percepción de la Altura Tonal , Estimulación Acústica , Discriminación de la Altura Tonal , Inteligibilidad del Habla
4.
Psychophysiology ; 60(11): e14362, 2023 Nov.
Artículo en Inglés | MEDLINE | ID: mdl-37350379

RESUMEN

The most prominent acoustic features in speech are intensity modulations, represented by the amplitude envelope of speech. Synchronization of neural activity with these modulations supports speech comprehension. As the acoustic modulation of speech is related to the production of syllables, investigations of neural speech tracking commonly do not distinguish between lower-level acoustic (envelope modulation) and higher-level linguistic (syllable rate) information. Here we manipulated speech intelligibility using noise-vocoded speech and investigated the spectral dynamics of neural speech processing, across two studies at cortical and subcortical levels of the auditory hierarchy, using magnetoencephalography. Overall, cortical regions mostly track the syllable rate, whereas subcortical regions track the acoustic envelope. Furthermore, with less intelligible speech, tracking of the modulation rate becomes more dominant. Our study highlights the importance of distinguishing between envelope modulation and syllable rate and provides novel possibilities to better understand differences between auditory processing and speech/language processing disorders.


Asunto(s)
Percepción del Habla , Habla , Humanos , Magnetoencefalografía , Ruido , Cognición , Estimulación Acústica , Inteligibilidad del Habla
5.
Artículo en Inglés | MEDLINE | ID: mdl-37159306

RESUMEN

Perception with electric neuroprostheses is sometimes expected to be simulated using properly designed physical stimuli. Here, we examined a new acoustic vocoder model for electric hearing with cochlear implants (CIs) and hypothesized that comparable speech encoding can lead to comparable perceptual patterns for CI and normal hearing (NH) listeners. Speech signals were encoded using FFT-based signal processing stages including band-pass filtering, temporal envelope extraction, maxima selection, and amplitude compression and quantization. These stages were specifically implemented in the same manner by an Advanced Combination Encoder (ACE) strategy in CI processors and Gaussian-enveloped Tones (GET) or Noise (GEN) vocoders for NH. Adaptive speech reception thresholds (SRTs) in noise were measured using four Mandarin sentence corpora. Initial consonant (11 monosyllables) and final vowel (20 monosyllables) recognition were also measured. NaÏve NH listeners were tested using vocoded speech with the proposed GET/GEN vocoders as well as conventional vocoders (controls). Experienced CI listeners were tested using their daily-used processors. Results showed that: 1) there was a significant training effect on GET vocoded speech perception; 2) the GEN vocoded scores (SRTs with four corpora and consonant and vowel recognition scores) as well as the phoneme-level confusion pattern matched with the CI scores better than controls. The findings suggest that the same signal encoding implementations may lead to similar perceptual patterns simultaneously in multiple perception tasks. This study highlights the importance of faithfully replicating all signal processing stages in the modeling of perceptual patterns in sensory neuroprostheses. This approach has the potential to enhance our understanding of CI perception and accelerate the engineering of prosthetic interventions. The GET/GEN MATLAB program is freely available athttps://github.com/BetterCI/GETVocoder.


Asunto(s)
Implantes Cocleares , Percepción del Habla , Humanos , Audición , Acústica , Inteligibilidad del Habla , Estimulación Acústica
6.
Neuroimage ; 272: 120040, 2023 05 15.
Artículo en Inglés | MEDLINE | ID: mdl-36935084

RESUMEN

During listening, brain activity tracks the rhythmic structures of speech signals. Here, we directly dissociated the contribution of neural envelope tracking in the processing of speech acoustic cues from that related to linguistic processing. We examined the neural changes associated with the comprehension of Noise-Vocoded (NV) speech using magnetoencephalography (MEG). Participants listened to NV sentences in a 3-phase training paradigm: (1) pre-training, where NV stimuli were barely comprehended, (2) training with exposure of the original clear version of speech stimulus, and (3) post-training, where the same stimuli gained intelligibility from the training phase. Using this paradigm, we tested if the neural responses of a speech signal was modulated by its intelligibility without any change in its acoustic structure. To test the influence of spectral degradation on neural envelope tracking independently of training, participants listened to two types of NV sentences (4-band and 2-band NV speech), but were only trained to understand 4-band NV speech. Significant changes in neural tracking were observed in the delta range in relation to the acoustic degradation of speech. However, we failed to find a direct effect of intelligibility on the neural tracking of speech envelope in both theta and delta ranges, in both auditory regions-of-interest and whole-brain sensor-space analyses. This suggests that acoustics greatly influence the neural tracking response to speech envelope, and that caution needs to be taken when choosing the control signals for speech-brain tracking analyses, considering that a slight change in acoustic parameters can have strong effects on the neural tracking response.


Asunto(s)
Percepción del Habla , Habla , Humanos , Habla/fisiología , Estimulación Acústica , Percepción del Habla/fisiología , Magnetoencefalografía , Ruido , Inteligibilidad del Habla
7.
Medicine (Baltimore) ; 102(6): e32900, 2023 Feb 10.
Artículo en Inglés | MEDLINE | ID: mdl-36820601

RESUMEN

BACKGROUND: Benefits of intensive speech treatment have been documented for a range of speech signs in English speakers with Parkinson's Disease (PD). However, the answer to a critical question that whether the same treatment benefits speech variables including intelligibility in Mandarin speakers is still unclear. In order to develop a targeted speech treatment for Mandarin speakers with PD, we reviewed the efficacy of intensive speech treatment to improve vocal loudness and functional communication and discuss possible explanations for efficacy on Mandarin speakers with PD. METHODS: Literatures about intensive speech treatment for Mandarin speakers with PD were retrieved from PubMed, Web of Science, Embase, China National Knowledge Infrastructure (CNKI), Wanfang and Weipu Database for Chinese Technical Periodicals (VIP) Database. Search strategy was (voice therapy OR speech therapy OR voice treatment OR speech treatment OR voice training OR speech training OR voice rehabilitation OR speech rehabilitation OR Lee Silverman voice treatment OR intensive speech treatment) and (Parkinson's disease) and (Mandarin speakers OR Chinese OR Chinese people). RESULTS: Five randomized controlled trials were selected and possible explanations for efficacy on individuals with PD are discussed. Further research directions are suggested. CONCLUSION: The existing evidence from treatment efficacy studies of intensive speech treatment provides support for improving vocal loudness, speech intelligibility, pitch and rate in Mandarin speakers with PD. Our future research will continue to work to conduct a large sample multicenter randomized controlled trial to provide high quality evidence and understand the basic mechanisms accompanying treatment-related change.


Asunto(s)
Enfermedad de Parkinson , Trastornos de la Voz , Voz , Humanos , Habla , Enfermedad de Parkinson/diagnóstico , Inteligibilidad del Habla , Logopedia , Estudios Multicéntricos como Asunto
8.
Cereb Cortex ; 33(3): 691-708, 2023 01 05.
Artículo en Inglés | MEDLINE | ID: mdl-35253871

RESUMEN

Online speech processing imposes significant computational demands on the listening brain, the underlying mechanisms of which remain poorly understood. Here, we exploit the perceptual "pop-out" phenomenon (i.e. the dramatic improvement of speech intelligibility after receiving information about speech content) to investigate the neurophysiological effects of prior expectations on degraded speech comprehension. We recorded electroencephalography (EEG) and pupillometry from 21 adults while they rated the clarity of noise-vocoded and sine-wave synthesized sentences. Pop-out was reliably elicited following visual presentation of the corresponding written sentence, but not following incongruent or neutral text. Pop-out was associated with improved reconstruction of the acoustic stimulus envelope from low-frequency EEG activity, implying that improvements in perceptual clarity were mediated via top-down signals that enhanced the quality of cortical speech representations. Spectral analysis further revealed that pop-out was accompanied by a reduction in theta-band power, consistent with predictive coding accounts of acoustic filling-in and incremental sentence processing. Moreover, delta-band power, alpha-band power, and pupil diameter were all increased following the provision of any written sentence information, irrespective of content. Together, these findings reveal distinctive profiles of neurophysiological activity that differentiate the content-specific processes associated with degraded speech comprehension from the context-specific processes invoked under adverse listening conditions.


Asunto(s)
Motivación , Percepción del Habla , Ruido , Electroencefalografía , Estimulación Acústica , Inteligibilidad del Habla/fisiología , Percepción del Habla/fisiología
9.
Cereb Cortex ; 33(5): 2215-2228, 2023 02 20.
Artículo en Inglés | MEDLINE | ID: mdl-35695785

RESUMEN

The envelope is essential for speech perception. Recent studies have shown that cortical activity can track the acoustic envelope. However, whether the tracking strength reflects the extent of speech intelligibility processing remains controversial. Here, using stereo-electroencephalogram technology, we directly recorded the activity in human auditory cortex while subjects listened to either natural or noise-vocoded speech. These 2 stimuli have approximately identical envelopes, but the noise-vocoded speech does not have speech intelligibility. According to the tracking lags, we revealed 2 stages of envelope tracking: an early high-γ (60-140 Hz) power stage that preferred the noise-vocoded speech and a late θ (4-8 Hz) phase stage that preferred the natural speech. Furthermore, the decoding performance of high-γ power was better in primary auditory cortex than in nonprimary auditory cortex, consistent with its short tracking delay, while θ phase showed better decoding performance in right auditory cortex. In addition, high-γ responses with sustained temporal profiles in nonprimary auditory cortex were dominant in both envelope tracking and decoding. In sum, we suggested a functional dissociation between high-γ power and θ phase: the former reflects fast and automatic processing of brief acoustic features, while the latter correlates to slow build-up processing facilitated by speech intelligibility.


Asunto(s)
Corteza Auditiva , Percepción del Habla , Humanos , Habla/fisiología , Corteza Auditiva/fisiología , Inteligibilidad del Habla , Estimulación Acústica , Electroencefalografía , Percepción del Habla/fisiología
10.
Hear Res ; 426: 108607, 2022 12.
Artículo en Inglés | MEDLINE | ID: mdl-36137861

RESUMEN

When a person listens to sound, the brain time-locks to specific aspects of the sound. This is called neural tracking and it can be investigated by analysing neural responses (e.g., measured by electroencephalography) to continuous natural speech. Measures of neural tracking allow for an objective investigation of a range of auditory and linguistic processes in the brain during natural speech perception. This approach is more ecologically valid than traditional auditory evoked responses and has great potential for research and clinical applications. This article reviews the neural tracking framework and highlights three prominent examples of neural tracking analyses: neural tracking of the fundamental frequency of the voice (f0), the speech envelope and linguistic features. Each of these analyses provides a unique point of view into the human brain's hierarchical stages of speech processing. F0-tracking assesses the encoding of fine temporal information in the early stages of the auditory pathway, i.e., from the auditory periphery up to early processing in the primary auditory cortex. Envelope tracking reflects bottom-up and top-down speech-related processes in the auditory cortex and is likely necessary but not sufficient for speech intelligibility. Linguistic feature tracking (e.g. word or phoneme surprisal) relates to neural processes more directly related to speech intelligibility. Together these analyses form a multi-faceted objective assessment of an individual's auditory and linguistic processing.


Asunto(s)
Corteza Auditiva , Percepción del Habla , Humanos , Vías Auditivas , Estimulación Acústica , Percepción del Habla/fisiología , Inteligibilidad del Habla , Corteza Auditiva/fisiología , Electroencefalografía
11.
Am J Speech Lang Pathol ; 31(3): 1354-1367, 2022 05 10.
Artículo en Inglés | MEDLINE | ID: mdl-35394803

RESUMEN

PURPOSE: This study investigated the effects of intensive voice treatment on subjective and objective measures of speech production in Mandarin speakers with hypokinetic dysarthria. METHOD: Nine Mandarin speakers with hypokinetic dysarthria due to Parkinson's disease received 4 weeks of intensive voice treatment (4 × 60 min per week). The speakers were recorded reading a passage before treatment (PRE), immediately after treatment (POST), and at 6-month follow-up (FU). Listeners (n = 15) rated relative ease of understanding (EOU) of paired speech samples on a visual analogue scale. Acoustic analyses were performed. Changes in EOU, vocal intensity, global and local fundamental frequency (f o) variation, speech rate, and acoustic vowel space area (VSA) were examined. RESULTS: Increases were found in EOU and vocal intensity from PRE to POST and from PRE to FU, with no change found from POST to FU. Speech rate increased from PRE to POST, with limited evidence of an increase from PRE to FU and no change from POST to FU. No changes in global or local f o variation or in VSA were found. CONCLUSIONS: Intensive voice treatment shows promise for improving speech production in Mandarin speakers with hypokinetic dysarthria. Vocal intensity, speech rate, and, crucially, intelligibility, may improve for up to 6 months posttreatment. In contrast, f o variation and VSA may not increase following the treatment. SUPPLEMENTAL MATERIAL: https://doi.org/10.23641/asha.19529017.


Asunto(s)
Disartria , Enfermedad de Parkinson , Acústica , Disartria/diagnóstico , Disartria/etiología , Disartria/terapia , Humanos , Enfermedad de Parkinson/complicaciones , Acústica del Lenguaje , Inteligibilidad del Habla , Medición de la Producción del Habla
12.
PLoS One ; 17(4): e0267588, 2022.
Artículo en Inglés | MEDLINE | ID: mdl-35468160

RESUMEN

The present study aimed to investigate the effects of degraded speech perception and binaural unmasking using functional near-infrared spectroscopy (fNIRS). Normal hearing listeners were tested when attending to unprocessed or vocoded speech, presented to the left ear at two speech-to-noise ratios (SNRs). Additionally, by comparing monaural versus diotic masker noise, we measured binaural unmasking. Our primary research question was whether the prefrontal cortex and temporal cortex responded differently to varying listening configurations. Our a priori regions of interest (ROIs) were located at the left dorsolateral prefrontal cortex (DLPFC) and auditory cortex (AC). The left DLPFC has been reported to be involved in attentional processes when listening to degraded speech and in spatial hearing processing, while the AC has been reported to be sensitive to speech intelligibility. Comparisons of cortical activity between these two ROIs revealed significantly different fNIRS response patterns. Further, we showed a significant and positive correlation between self-reported task difficulty levels and fNIRS responses in the DLPFC, with a negative but non-significant correlation for the left AC, suggesting that the two ROIs played different roles in effortful speech perception. Our secondary question was whether activity within three sub-regions of the lateral PFC (LPFC) including the DLPFC was differentially affected by varying speech-noise configurations. We found significant effects of spectral degradation and SNR, and significant differences in fNIRS response amplitudes between the three regions, but no significant interaction between ROI and speech type, or between ROI and SNR. When attending to speech with monaural and diotic noises, participants reported the latter conditions being easier; however, no significant main effect of masker condition on cortical activity was observed. For cortical responses in the LPFC, a significant interaction between SNR and masker condition was observed. These findings suggest that binaural unmasking affects cortical activity through improving speech reception threshold in noise, rather than by reducing effort exerted.


Asunto(s)
Espectroscopía Infrarroja Corta , Percepción del Habla , Estimulación Acústica/métodos , Humanos , Ruido , Inteligibilidad del Habla , Percepción del Habla/fisiología
13.
J Neural Eng ; 19(2)2022 04 12.
Artículo en Inglés | MEDLINE | ID: mdl-35234663

RESUMEN

Objectives. Electroencephalography (EEG) can be used to decode selective attention in cochlear implant (CI) users. This work investigates if selective attention to an attended speech source in the presence of a concurrent speech source can predict speech understanding in CI users.Approach. CI users were instructed to attend to one out of two speech streams while EEG was recorded. Both speech streams were presented to the same ear and at different signal to interference ratios (SIRs). Speech envelope reconstruction of the to-be-attended speech from EEG was obtained by training decoders using regularized least squares. The correlation coefficient between the reconstructed and the attended (ρASIR)or the unattendedρUSIR speech stream at each SIR was computed. Additionally, we computed the difference correlation coefficient at the same(ρDiff= ρASIR-ρUSIR)and opposite SIR (ρDiffOpp= ρASIR-ρU-SIR).ρDiffcompares the attended and unattended correlation coefficient to speech sources presented at different presentation levels depending on SIR. In contrast,ρDiffOppcompares the attended and unattended correlation coefficients to speech sources presented at the same presentation level irrespective of SIR.Main results. Selective attention decoding in CI users is possible even if both speech streams are presented monaurally. A significant effect of SIR onρASIR,ρDiffandρDiffOpp, but not onρUSIR, was observed. Finally, the results show a significant correlation between speech understanding performance andρASIRas well as withρUSIRacross subjects. Moreover,ρDiffOppwhich is less affected by the CI artifact, also demonstrated a significant correlation with speech understanding.Significance. Selective attention decoding in CI users is possible, however care needs to be taken with the CI artifact and the speech material used to train the decoders. These results are important for future development of objective speech understanding measures for CI users.


Asunto(s)
Implantación Coclear , Implantes Cocleares , Percepción del Habla , Estimulación Acústica/métodos , Electroencefalografía/métodos , Humanos , Inteligibilidad del Habla
14.
Altern Ther Health Med ; 28(3): 48-51, 2022 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-35180096

RESUMEN

OBJECTIVE: To determine the effectiveness of Kinesio Taping (KT) and Manipulation Therapy (MT) on drooling and speech intelligibility in children with oral dysphagia. METHODS: A randomized clinical trial was conducted at Helping Hand Institute of Rehabilitation Sciences in Mansehra, Pakistan. A total of 20 patients were recruited via the random sampling technique and later assigned to one of two groups: KT (n = 10) or MT (n = 10). Every patient in both groups received their intervention 5 days a week for 1 month (20 sessions total), and each session lasted 45 minutes. Data was collected and analyzed at baseline and 1 month. Drooling was assessed via the Modified Teachers' Drooling Scale and speech intelligibility determined via the 7-Point Intelligibility Rating Scale. Mann Whitney U-test was used for between-group comparisons and for within-group comparisons the Wilcoxon signed-rank test and their effect size was used. RESULTS: The mean age of study patients was 5.4 years. Of the 20 patients, 14 were male and 6 were female. Within-group comparisons showed significant improvement in both drooling and speech intelligibility (P < .05), while between-group comparisons showed no significant difference (P ≥ .05) in either the KT or MT group regarding drooling severity and speech intelligibility. CONCLUSION: KT and MT significantly improved drooling and speech intelligibility.


Asunto(s)
Trastornos de Deglución , Manipulaciones Musculoesqueléticas , Sialorrea , Niño , Preescolar , Trastornos de Deglución/terapia , Femenino , Humanos , Masculino , Proyectos Piloto , Sialorrea/terapia , Inteligibilidad del Habla
15.
Int J Audiol ; 61(1): 1-11, 2022 Jan.
Artículo en Inglés | MEDLINE | ID: mdl-34154488

RESUMEN

OBJECTIVE: Telecommunication can be difficult in the presence of noise or hearing loss. The purpose of this study was to systematically review evidence regarding the effects of text supplementation (e.g. captions, subtitles) of auditory or auditory-visual signals on speech intelligibility for listeners with normal or impaired hearing. DESIGN: Three databases were searched. Articles were evaluated for inclusion based on the Population Intervention Comparison Outcome framework. The Effective Public Health Practice Project instrument was used to evaluate the quality of the identified articles. STUDY SAMPLE: After duplicates were removed, the titles and abstracts of 2019 articles were screened. Forty-six full texts were reviewed; ten met inclusion criteria. RESULTS: The quality of all ten articles was moderate or strong. The articles demonstrated that text added to auditory (or auditory-visual) signals improved speech intelligibility and that the benefits were largest when auditory signal integrity was low, accuracy of the text was high, and the auditory signal and text were synchronous. Age and hearing loss did not affect benefits from the addition of text. CONCLUSIONS: Although only based on ten studies, these data support the use of text as a supplement during telecommunication, such as while watching television or during telehealth appointments.


Asunto(s)
Sordera , Pérdida Auditiva Sensorineural , Pérdida Auditiva , Percepción del Habla , Telecomunicaciones , Umbral Auditivo , Suplementos Dietéticos , Audición , Pérdida Auditiva/diagnóstico , Humanos , Inteligibilidad del Habla
16.
J Neural Eng ; 18(6)2021 11 15.
Artículo en Inglés | MEDLINE | ID: mdl-34706347

RESUMEN

Objective.Currently, only behavioral speech understanding tests are available, which require active participation of the person being tested. As this is infeasible for certain populations, an objective measure of speech intelligibility is required. Recently, brain imaging data has been used to establish a relationship between stimulus and brain response. Linear models have been successfully linked to speech intelligibility but require per-subject training. We present a deep-learning-based model incorporating dilated convolutions that operates in a match/mismatch paradigm. The accuracy of the model's match/mismatch predictions can be used as a proxy for speech intelligibility without subject-specific (re)training.Approach.We evaluated the performance of the model as a function of input segment length, electroencephalography (EEG) frequency band and receptive field size while comparing it to multiple baseline models. Next, we evaluated performance on held-out data and finetuning. Finally, we established a link between the accuracy of our model and the state-of-the-art behavioral MATRIX test.Main results.The dilated convolutional model significantly outperformed the baseline models for every input segment length, for all EEG frequency bands except the delta and theta band, and receptive field sizes between 250 and 500 ms. Additionally, finetuning significantly increased the accuracy on a held-out dataset. Finally, a significant correlation (r= 0.59,p= 0.0154) was found between the speech reception threshold (SRT) estimated using the behavioral MATRIX test and our objective method.Significance.Our method is the first to predict the SRT from EEG for unseen subjects, contributing to objective measures of speech intelligibility.


Asunto(s)
Inteligibilidad del Habla , Percepción del Habla , Estimulación Acústica , Encéfalo , Electroencefalografía/métodos , Audición/fisiología , Humanos , Inteligibilidad del Habla/fisiología , Percepción del Habla/fisiología
17.
J Assoc Res Otolaryngol ; 22(6): 703-717, 2021 12.
Artículo en Inglés | MEDLINE | ID: mdl-34581879

RESUMEN

Whilst functional neuroimaging has been used to investigate cortical processing of degraded speech in adults, much less is known about how these signals are processed in children. An enhanced understanding of cortical correlates of poor speech perception in children would be highly valuable to oral communication applications, including hearing devices. We utilised vocoded speech stimuli to investigate brain responses to degraded speech in 29 normally hearing children aged 6-12 years. Intelligibility of the speech stimuli was altered in two ways by (i) reducing the number of spectral channels and (ii) reducing the amplitude modulation depth of the signal. A total of five different noise-vocoded conditions (with zero, partial or high intelligibility) were presented in an event-related format whilst participants underwent functional near-infrared spectroscopy (fNIRS) neuroimaging. Participants completed a word recognition task during imaging, as well as a separate behavioural speech perception assessment. fNIRS recordings revealed statistically significant sensitivity to stimulus intelligibility across several brain regions. More intelligible stimuli elicited stronger responses in temporal regions, predominantly within the left hemisphere, while right inferior parietal regions showed an opposite, negative relationship. Although there was some evidence that partially intelligible stimuli elicited the strongest responses in the left inferior frontal cortex, a region previous studies have suggested is associated with effortful listening in adults, this effect did not reach statistical significance. These results further our understanding of cortical mechanisms underlying successful speech perception in children. Furthermore, fNIRS holds promise as a clinical technique to help assess speech intelligibility in paediatric populations.


Asunto(s)
Percepción Auditiva , Espectroscopía Infrarroja Corta/métodos , Inteligibilidad del Habla , Percepción del Habla , Estimulación Acústica/métodos , Percepción Auditiva/fisiología , Niño , Femenino , Audición/fisiología , Humanos , Lenguaje , Masculino , Neuroimagen , Ruido , Percepción del Habla/fisiología
18.
J Speech Lang Hear Res ; 64(9): 3697-3706, 2021 09 14.
Artículo en Inglés | MEDLINE | ID: mdl-34403278

RESUMEN

Purpose Listeners shift their listening strategies between lower level acoustic information and higher level semantic information to prioritize maximum speech intelligibility in challenging listening conditions. Although increasing task demands via acoustic degradation modulates lexical-semantic processing, the neural mechanisms underlying different listening strategies are unclear. The current study examined the extent to which encoding of lower level acoustic cues is modulated by task demand and associations with lexical-semantic processes. Method Electroencephalography was acquired while participants listened to sentences in the presence of four-talker babble that contained either higher or lower probability final words. Task difficulty was modulated by time available to process responses. Cortical tracking of speech-neural correlates of acoustic temporal envelope processing-were estimated using temporal response functions. Results Task difficulty did not affect cortical tracking of temporal envelope of speech under challenging listening conditions. Neural indices of lexical-semantic processing (N400 amplitudes) were larger with increased task difficulty. No correlations were observed between the cortical tracking of temporal envelope of speech and lexical-semantic processes, even after controlling for the effect of individualized signal-to-noise ratios. Conclusions Cortical tracking of the temporal envelope of speech and semantic processing are differentially influenced by task difficulty. While increased task demands modulated higher level semantic processing, cortical tracking of the temporal envelope of speech may be influenced by task difficulty primarily when the demand is manipulated in terms of acoustic properties of the stimulus, consistent with an emerging perspective in speech perception.


Asunto(s)
Semántica , Percepción del Habla , Estimulación Acústica , Acústica , Electroencefalografía , Potenciales Evocados , Femenino , Humanos , Masculino , Inteligibilidad del Habla
19.
Neuroimage ; 240: 118385, 2021 10 15.
Artículo en Inglés | MEDLINE | ID: mdl-34256138

RESUMEN

In this study we used functional near-infrared spectroscopy (fNIRS) to investigate neural responses in normal-hearing adults as a function of speech recognition accuracy, intelligibility of the speech stimulus, and the manner in which speech is distorted. Participants listened to sentences and reported aloud what they heard. Speech quality was distorted artificially by vocoding (simulated cochlear implant speech) or naturally by adding background noise. Each type of distortion included high and low-intelligibility conditions. Sentences in quiet were used as baseline comparison. fNIRS data were analyzed using a newly developed image reconstruction approach. First, elevated cortical responses in the middle temporal gyrus (MTG) and middle frontal gyrus (MFG) were associated with speech recognition during the low-intelligibility conditions. Second, activation in the MTG was associated with recognition of vocoded speech with low intelligibility, whereas MFG activity was largely driven by recognition of speech in background noise, suggesting that the cortical response varies as a function of distortion type. Lastly, an accuracy effect in the MFG demonstrated significantly higher activation during correct perception relative to incorrect perception of speech. These results suggest that normal-hearing adults (i.e., untrained listeners of vocoded stimuli) do not exploit the same attentional mechanisms of the frontal cortex used to resolve naturally degraded speech and may instead rely on segmental and phonetic analyses in the temporal lobe to discriminate vocoded speech.


Asunto(s)
Estimulación Acústica/métodos , Implantes Cocleares , Lóbulo Frontal/fisiología , Inteligibilidad del Habla/fisiología , Percepción del Habla/fisiología , Lóbulo Temporal/fisiología , Adolescente , Adulto , Femenino , Lóbulo Frontal/diagnóstico por imagen , Humanos , Masculino , Ruido/efectos adversos , Espectroscopía Infrarroja Corta/métodos , Lóbulo Temporal/diagnóstico por imagen , Adulto Joven
20.
Sci Rep ; 11(1): 15117, 2021 07 23.
Artículo en Inglés | MEDLINE | ID: mdl-34302032

RESUMEN

Our acoustic environment contains a plethora of complex sounds that are often in motion. To gauge approaching danger and communicate effectively, listeners need to localize and identify sounds, which includes determining sound motion. This study addresses which acoustic cues impact listeners' ability to determine sound motion. Signal envelope (ENV) cues are implicated in both sound motion tracking and stimulus intelligibility, suggesting that these processes could be competing for sound processing resources. We created auditory chimaera from speech and noise stimuli and varied the number of frequency bands, effectively manipulating speech intelligibility. Normal-hearing adults were presented with stationary or moving chimaeras and reported perceived sound motion and content. Results show that sensitivity to sound motion is not affected by speech intelligibility, but shows a clear difference for original noise and speech stimuli. Further, acoustic chimaera with speech-like ENVs which had intelligible content induced a strong bias in listeners to report sounds as stationary. Increasing stimulus intelligibility systematically increased that bias and removing intelligible content reduced it, suggesting that sound content may be prioritized over sound motion. These findings suggest that sound motion processing in the auditory system can be biased by acoustic parameters related to speech intelligibility.


Asunto(s)
Percepción Auditiva/fisiología , Percepción de Movimiento/fisiología , Inteligibilidad del Habla/fisiología , Estimulación Acústica/métodos , Adulto , Umbral Auditivo/fisiología , Señales (Psicología) , Femenino , Audición/fisiología , Pruebas Auditivas/métodos , Humanos , Masculino , Movimiento (Física) , Ruido , Enmascaramiento Perceptual/fisiología , Sonido , Acústica del Lenguaje , Percepción del Habla/fisiología , Adulto Joven
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA