Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 36
Filtrar
1.
Ear Hear ; 2024 Jun 18.
Artigo em Inglês | MEDLINE | ID: mdl-38886880

RESUMO

OBJECTIVES: In clinical and laboratory settings, speech recognition is typically assessed in a way that cannot distinguish accurate auditory perception from misperception that was mentally repaired or inferred from context. Previous work showed that the process of repairing misperceptions elicits greater listening effort, and that this elevated effort lingers well after the sentence is heard. That result suggests that cognitive repair strategies might appear successful when testing a single utterance but fail for everyday continuous conversational speech. The present study tested the hypothesis that the effort of repairing misperceptions has the consequence of carrying over to interfere with perception of later words after the sentence. DESIGN: Stimuli were open-set coherent sentences that were presented intact or with a word early in the sentence replaced with noise, forcing the listener to use later context to mentally repair the missing word. Sentences were immediately followed by digit triplets, which served to probe carryover effort from the sentence. Control conditions allowed for the comparison to intact sentences that did not demand mental repair, as well as to listening conditions that removed the need to attend to the post-sentence stimuli, or removed the post-sentence digits altogether. Intelligibility scores for the sentences and digits were accompanied by time-series measurements of pupil dilation to assess cognitive load during the task, as well as subjective rating of effort. Participants included adults with cochlear implants (CIs), as well as an age-matched group and a younger group of listeners with typical hearing for comparison. RESULTS: For the CI group, needing to repair a missing word during a sentence resulted in more errors on the digits after the sentence, especially when the repair process did not result in a coherent sensible perception. Sentences that needed repair also contained more errors on the words that were unmasked. All groups showed substantial increase of pupil dilation when sentences required repair, even when the repair was successful. Younger typical hearing listeners showed clear differences in moment-to-moment allocation of effort in the different conditions, while the other groups did not. CONCLUSIONS: For CI listeners, the effort of needing to repair misperceptions in a sentence can last long enough to interfere with words that follow the sentence. This pattern could pose a serious problem for regular communication but would go overlooked in typical testing with single utterances, where a listener has a chance to repair misperceptions before responding. Carryover effort was not predictable by basic intelligibility scores, but can be revealed in behavioral data when sentences are followed immediately by extra probe words such as digits.

2.
Ear Hear ; 45(4): 915-928, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38389129

RESUMO

OBJECTIVES: Understanding speech in noise (SIN) is the dominant complaint of individuals with hearing loss. For decades, the default test of speech perception in routine audiologic assessment has been monosyllabic word recognition in quiet (WRQ), which does not directly address patient concerns, leading some to advocate that measures of SIN should be integrated into routine practice. However, very little is known with regard to how SIN abilities are affected by different types of hearing loss. Here, we examine performance on clinical measures of WRQ and SIN in a large patient base consisting of a variety of hearing loss types, including conductive (CHL), mixed (MHL), and sensorineural (SNHL) losses. DESIGN: In a retrospective study, we examined data from 5593 patients (51% female) who underwent audiometric assessment at the Stanford Ear Institute. All individuals completed pure-tone audiometry, and speech perception testing of monaural WRQ, and monaural QuickSIN. Patient ages ranged from 18 to 104 years (average = 57). The average age in years for the different classifications of hearing loss was 51.1 (NH), 48.5 (CHL), 64.2 (MHL), and 68.5 (SNHL), respectively. Generalized linear mixed-effect models and quartile regression were used to determine the relationship between hearing loss type and severity for the different speech-recognition outcome measures. RESULTS: Patients with CHL had similar performance to patients with normal hearing on both WRQ and QuickSIN, regardless of the hearing loss severity. In patients with MHL or SNHL, WRQ scores remained largely excellent with increasing hearing loss until the loss was moderately severe or worse. In contrast, QuickSIN signal to noise ratio (SNR) losses showed an orderly systematic decrease as the degree of hearing loss became more severe. This effect scaled with the data, with threshold-QuickSIN relationships absent for CHL, and becoming increasingly stronger for MHL and strongest in patients with SNHL. However, the variability in these data suggests that only 57% of the variance in WRQ scores, and 50% of the variance in QuickSIN SNR losses, could be accounted for by the audiometric thresholds. Patients who would not be differentiated by WRQ scores are shown to be potentially differentiable by SIN scores. CONCLUSIONS: In this data set, conductive hearing loss had little effect on WRQ scores or QuickSIN SNR losses. However, for patients with MHL or SNHL, speech perception abilities decreased as the severity of the hearing loss increased. In these data, QuickSIN SNR losses showed deficits in performance with degrees of hearing loss that yielded largely excellent WRQ scores. However, the considerable variability in the data suggests that even after classifying patients according to their type of hearing loss, hearing thresholds only account for a portion of the variance in speech perception abilities, particularly in noise. These results are consistent with the idea that variables such as cochlear health and aging add explanatory power over audibility alone.


Assuntos
Audiometria de Tons Puros , Perda Auditiva Neurossensorial , Ruído , Percepção da Fala , Humanos , Feminino , Pessoa de Meia-Idade , Masculino , Idoso , Adulto , Estudos Retrospectivos , Idoso de 80 Anos ou mais , Adulto Jovem , Adolescente , Perda Auditiva Neurossensorial/fisiopatologia , Índice de Gravidade de Doença , Perda Auditiva Condutiva/fisiopatologia , Perda Auditiva Condutiva-Neurossensorial Mista/fisiopatologia , Perda Auditiva Condutiva-Neurossensorial Mista/reabilitação , Perda Auditiva/fisiopatologia
3.
J Acoust Soc Am ; 155(4): 2407-2437, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38568143

RESUMO

The channel vocoder has become a useful tool to understand the impact of specific forms of auditory degradation-particularly the spectral and temporal degradation that reflect cochlear-implant processing. Vocoders have many parameters that allow researchers to answer questions about cochlear-implant processing in ways that overcome some logistical complications of controlling for factors in individual cochlear implant users. However, there is such a large variety in the implementation of vocoders that the term "vocoder" is not specific enough to describe the signal processing used in these experiments. Misunderstanding vocoder parameters can result in experimental confounds or unexpected stimulus distortions. This paper highlights the signal processing parameters that should be specified when describing vocoder construction. The paper also provides guidance on how to determine vocoder parameters within perception experiments, given the experimenter's goals and research questions, to avoid common signal processing mistakes. Throughout, we will assume that experimenters are interested in vocoders with the specific goal of better understanding cochlear implants.


Assuntos
Implante Coclear , Implantes Cocleares , Processamento de Sinais Assistido por Computador
4.
J Acoust Soc Am ; 153(3): 1623, 2023 03.
Artigo em Inglês | MEDLINE | ID: mdl-37002094

RESUMO

The papers in this special issue provide a critical look at some historical ideas that have had an influence on research and teaching in the field of speech communication. They also address widely used methodologies or address long-standing methodological challenges in the areas of speech perception and speech production. The goal is to reconsider and evaluate the need for caution or replacement of historical ideas with more modern results and methods. The contributions provide respectful historical context to the classic ideas, as well as new original research or discussion that clarifies the limitations of the original ideas.


Assuntos
Percepção da Fala , Fala , Comunicação
5.
Ear Hear ; 43(3): 764-772, 2022.
Artigo em Inglês | MEDLINE | ID: mdl-34966157

RESUMO

The spectral ripple discrimination task is a psychophysical measure that has been found to correlate with speech recognition in listeners with cochlear implants (CIs). However, at ripple densities above a critical value (around 2 RPO, but device-specific), the sparse spectral sampling of CI processors results in stimulus distortions resulting in aliasing and unintended changes in modulation depth. As a result, spectral ripple thresholds above a certain number are not ordered monotonically along the RPO dimension and thus cannot be considered better or worse spectral resolution than each other, thus undermining correlation measurements. These stimulus distortions are not remediated by changing stimulus phase, indicating these issues cannot be solved by spectrotemporally modulated stimuli. Speech generally has very low-density spectral modulations, leading to questions about the mechanism of correlation between high ripple thresholds and speech recognition. Existing data showing correlations between ripple discrimination and speech recognition include many observations above the aliasing limit. These scores should be treated with caution, and experimenters could benefit by prospectively considering the limitations of the spectral ripple test.


Assuntos
Implante Coclear , Implantes Cocleares , Percepção da Fala , Humanos , Fala
6.
J Acoust Soc Am ; 152(3): 1300, 2022 09.
Artigo em Inglês | MEDLINE | ID: mdl-36182279

RESUMO

Perception of word stress is an important aspect of recognizing speech, guiding the listener toward candidate words based on the perceived stress pattern. Cochlear implant (CI) signal processing is likely to disrupt some of the available cues for word stress, particularly vowel quality and pitch contour changes. In this study, we used a cue weighting paradigm to investigate differences in stress cue weighting patterns between participants listening with CIs and those with normal hearing (NH). We found that participants with CIs gave less weight to frequency-based pitch and vowel quality cues than NH listeners but compensated by upweighting vowel duration and intensity cues. Nonetheless, CI listeners' stress judgments were also significantly influenced by vowel quality and pitch, and they modulated their usage of these cues depending on the specific word pair in a manner similar to NH participants. In a series of separate online experiments with NH listeners, we simulated aspects of bimodal hearing by combining low-pass filtered speech with a vocoded signal. In these conditions, participants upweighted pitch and vowel quality cues relative to a fully vocoded control condition, suggesting that bimodal listening holds promise for restoring the stress cue weighting patterns exhibited by listeners with NH.


Assuntos
Implante Coclear , Implantes Cocleares , Percepção da Fala , Estimulação Acústica/métodos , Acústica , Sinais (Psicologia) , Audição , Humanos
7.
J Acoust Soc Am ; 152(3): 1394, 2022 09.
Artigo em Inglês | MEDLINE | ID: mdl-36182291

RESUMO

This paper examines some commonly used stimuli in speech perception experiments and raises questions about their use, or about the interpretations of previous results. The takeaway messages are: 1) the Hillenbrand vowels represent a particular dialect rather than a gold standard, and English vowels contain spectral dynamics that have been largely underappreciated, 2) the /ɑ/ context is very common but not clearly superior as a context for testing consonant perception, 3) /ɑ/ is particularly problematic when testing voice-onset-time perception because it introduces strong confounds in the formant transitions, 4) /dɑ/ is grossly overrepresented in neurophysiological studies and yet is insufficient as a generalized proxy for "speech perception," and 5) digit tests and matrix sentences including the coordinate response measure are systematically insensitive to important patterns in speech perception. Each of these stimulus sets and concepts is described with careful attention to their unique value and also cases where they might be misunderstood or over-interpreted.


Assuntos
Percepção da Fala , Voz , Idioma , Fonética , Fala , Acústica da Fala , Percepção da Fala/fisiologia
8.
Ear Hear ; 42(5): 1412-1427, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-33795617

RESUMO

OBJECTIVES: Cochlear implant (CI) recipients are at a severe disadvantage compared with normal-hearing listeners in distinguishing consonants that differ by place of articulation because the key relevant spectral differences are degraded by the implant. One component of that degradation is the upward shifting of spectral energy that occurs with a shallow insertion depth of a CI. The present study aimed to systematically measure the effects of spectral shifting on word recognition and phoneme categorization by specifically controlling the amount of shifting and using stimuli whose identification specifically depends on perceiving frequency cues. We hypothesized that listeners would be biased toward perceiving phonemes that contain higher-frequency components because of the upward frequency shift and that intelligibility would decrease as spectral shifting increased. DESIGN: Normal-hearing listeners (n = 15) heard sine wave-vocoded speech with simulated upward frequency shifts of 0, 2, 4, and 6 mm of cochlear space to simulate shallow CI insertion depth. Stimuli included monosyllabic words and /b/-/d/ and /∫/-/s/ continua that varied systematically by formant frequency transitions or frication noise spectral peaks, respectively. Recalibration to spectral shifting was operationally defined as shifting perceptual acoustic-phonetic mapping commensurate with the spectral shift. In other words, adjusting frequency expectations for both phonemes upward so that there is still a perceptual distinction, rather than hearing all upward-shifted phonemes as the higher-frequency member of the pair. RESULTS: For moderate amounts of spectral shifting, group data suggested a general "halfway" recalibration to spectral shifting, but individual data suggested a notably different conclusion: half of the listeners were able to recalibrate fully, while the other halves of the listeners were utterly unable to categorize shifted speech with any reliability. There were no participants who demonstrated a pattern intermediate to these two extremes. Intelligibility of words decreased with greater amounts of spectral shifting, also showing loose clusters of better- and poorer-performing listeners. Phonetic analysis of word errors revealed certain cues were more susceptible to being compromised due to a frequency shift (place and manner of articulation), while voicing was robust to spectral shifting. CONCLUSIONS: Shifting the frequency spectrum of speech has systematic effects that are in line with known properties of speech acoustics, but the ensuing difficulties cannot be predicted based on tonotopic mismatch alone. Difficulties are subject to substantial individual differences in the capacity to adjust acoustic-phonetic mapping. These results help to explain why speech recognition in CI listeners cannot be fully predicted by peripheral factors like electrode placement and spectral resolution; even among listeners with functionally equivalent auditory input, there is an additional factor of simply being able or unable to flexibly adjust acoustic-phonetic mapping. This individual variability could motivate precise treatment approaches guided by an individual's relative reliance on wideband frequency representation (even if it is mismatched) or limited frequency coverage whose tonotopy is preserved.


Assuntos
Implante Coclear , Implantes Cocleares , Percepção da Fala , Humanos , Reprodutibilidade dos Testes , Fala
9.
Ear Hear ; 42(3): 584-595, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-33002968

RESUMO

OBJECTIVES: Slowed speaking rate was examined for its effects on speech intelligibility, its interaction with the benefit of contextual cues, and the impact of these factors on listening effort in adults with cochlear implants. DESIGN: Participants (n = 21 cochlear implant users) heard high- and low-context sentences that were played at the original speaking rate, as well as a slowed (1.4× duration) speaking rate, using uniform pitch-synchronous time warping. In addition to intelligibility measures, changes in pupil dilation were measured as a time-varying index of processing load or listening effort. Slope of pupil size recovery to baseline after the sentence was used as an index of resolution of perceptual ambiguity. RESULTS: Speech intelligibility was better for high-context compared to low-context sentences and slightly better for slower compared to original-rate speech. Speech rate did not affect magnitude and latency of peak pupil dilation relative to sentence offset. However, baseline pupil size recovered more substantially for slower-rate sentences, suggesting easier processing in the moment after the sentence was over. The effect of slowing speech rate was comparable to changing a sentence from low context to high context. The effect of context on pupil dilation was not observed until after the sentence was over, and one of two analyses suggested that context had greater beneficial effects on listening effort when the speaking rate was slower. These patterns maintained even at perfect sentence intelligibility, suggesting that correct speech repetition does not guarantee efficient or effortless processing. With slower speaking rates, there was less variability in pupil dilation slopes following the sentence, implying mitigation of some of the difficulties shown by individual listeners who would otherwise demonstrate prolonged effort after a sentence is heard. CONCLUSIONS: Slowed speaking rate provides release from listening effort when hearing an utterance, particularly relieving effort that would have lingered after a sentence is over. Context arguably provides even more release from listening effort when speaking rate is slower. The pattern of prolonged pupil dilation for faster speech is consistent with increased need to mentally correct errors, although that exact interpretation cannot be verified with intelligibility data alone or with pupil data alone. A pattern of needing to dwell on a sentence to disambiguate misperceptions likely contributes to difficulty in running conversation where there are few opportunities to pause and resolve recently heard utterances.


Assuntos
Implante Coclear , Implantes Cocleares , Percepção da Fala , Adulto , Humanos , Ruído , Inteligibilidade da Fala
10.
J Acoust Soc Am ; 147(2): 852, 2020 02.
Artigo em Inglês | MEDLINE | ID: mdl-32113256

RESUMO

Voice onset time (VOT) is an acoustic property of stop consonants that is commonly manipulated in studies of phonetic perception. This paper contains a thorough description of the "progressive cutback and replacement" method of VOT manipulation, and comparison with other VOT manipulation techniques. Other acoustic properties that covary with VOT-such as fundamental frequency and formant transitions-are also discussed, along with considerations for testing VOT perception and its relationship to various other measures of auditory temporal or spectral processing. An implementation of the progressive cutback and replacement method in the Praat scripting language is presented, which is suitable for modifying natural speech for perceptual experiments involving VOT and/or related covarying F0 and intensity cues. Justifications are provided for the stimulus design choices and constraints implemented in the script.

11.
J Acoust Soc Am ; 147(1): 174, 2020 01.
Artigo em Inglês | MEDLINE | ID: mdl-32006986

RESUMO

Speech perception requires accommodation of a wide range of acoustic variability across talkers. A classic example is the perception of "sh" and "s" fricative sounds, which are categorized according to spectral details of the consonant itself, and also by the context of the voice producing it. Because women's and men's voices occupy different frequency ranges, a listener is required to make a corresponding adjustment of acoustic-phonetic category space for these phonemes when hearing different talkers. This pattern is commonplace in everyday speech communication, and yet might not be captured in accuracy scores for whole words, especially when word lists are spoken by a single talker. Phonetic accommodation for fricatives "s" and "sh" was measured in 20 cochlear implant (CI) users and in a variety of vocoder simulations, including those with noise carriers with and without peak picking, simulated spread of excitation, and pulsatile carriers. CI listeners showed strong phonetic accommodation as a group. Each vocoder produced phonetic accommodation except the 8-channel noise vocoder, despite its historically good match with CI users in word intelligibility. Phonetic accommodation is largely independent of linguistic factors and thus might offer information complementary to speech intelligibility tests which are partially affected by language processing.


Assuntos
Implantes Cocleares , Fonética , Percepção da Fala , Estimulação Acústica , Adolescente , Adulto , Humanos , Psicometria , Fatores Sexuais , Processamento de Sinais Assistido por Computador , Adulto Jovem
12.
J Acoust Soc Am ; 148(2): 496, 2020 08.
Artigo em Inglês | MEDLINE | ID: mdl-32873011

RESUMO

Listeners must accommodate acoustic differences between vocal tracts and speaking styles of conversation partners-a process called normalization or accommodation. This study explores what acoustic cues are used to make this perceptual adjustment by listeners with normal hearing or with cochlear implants, when the acoustic variability is related to the talker's gender. A continuum between /ʃ/ and /s/ was paired with naturally spoken vocalic contexts that were parametrically manipulated to vary by numerous cues for talker gender including fundamental frequency (F0), vocal tract length (formant spacing), and direct spectral contrast with the fricative. The goal was to examine relative contributions of these cues toward the tendency to have a lower-frequency acoustic boundary for fricatives spoken by men (found in numerous previous studies). Normal hearing listeners relied primarily on formant spacing and much less on F0. The CI listeners were individually variable, with the F0 cue emerging as the strongest cue on average.


Assuntos
Implantes Cocleares , Percepção da Fala , Estimulação Acústica , Acústica , Sinais (Psicologia) , Audição , Humanos , Masculino
13.
J Acoust Soc Am ; 147(5): 3626, 2020 05.
Artigo em Inglês | MEDLINE | ID: mdl-32486770

RESUMO

For cochlear-implant users with near-normal contralateral hearing, a mismatch between the frequency-to-place mapping in the two ears could produce a suboptimal performance. This study assesses tonotopic matches via binaural interactions. Dynamic interaural time-difference sensitivity was measured using bandpass-filtered pulse trains at different rates in the acoustic and implanted ear, creating binaural envelope beats. Sensitivity to beats should peak when the same tonotopic region is stimulated in both ears. All nine participants detected dynamic interaural timing differences and demonstrated some frequency selectivity. This method provides a guide to frequency-to-place mapping without compensation for inherent latency differences between the acoustic and implanted ears.


Assuntos
Implante Coclear , Implantes Cocleares , Surdez , Percepção da Fala , Estimulação Acústica , Surdez/diagnóstico , Testes Auditivos , Humanos
14.
Behav Res Methods ; 52(5): 2232-2255, 2020 10.
Artigo em Inglês | MEDLINE | ID: mdl-32291732

RESUMO

Eye-tracking is widely used throughout the scientific community, from vision science and psycholinguistics to marketing and human-computer interaction. Surprisingly, there is little consistency and transparency in preprocessing steps, making replicability and reproducibility difficult. To increase replicability, reproducibility, and transparency, a package in R (a free and widely used statistical programming environment) called gazeR was created to read and preprocess two types of data: gaze position and pupil size. For gaze position data, gazeR has functions for reading in raw eye-tracking data, formatting it for analysis, converting from gaze coordinates to areas of interest, and binning and aggregating data. For data from pupillometry studies, the gazeR package has functions for reading in and merging multiple raw pupil data files, removing observations with too much missing data, eliminating artifacts, blink identification and interpolation, subtractive baseline correction, and binning and aggregating data. The package is open-source and freely available for download and installation: https://github.com/dmirman/gazer . We provide step-by-step analyses of data from two tasks exemplifying the package's capabilities.


Assuntos
Piscadela , Pupila , Software , Humanos , Leitura , Reprodutibilidade dos Testes
15.
J Acoust Soc Am ; 146(5): 3373, 2019 11.
Artigo em Inglês | MEDLINE | ID: mdl-31795696

RESUMO

When hearing an ambiguous speech sound, listeners show a tendency to perceive it as a phoneme that would complete a real word, rather than completing a nonsense/fake word. For example, a sound that could be heard as either /b/ or /É¡/ is perceived as /b/ when followed by _ack but perceived as /É¡/ when followed by "_ap." Because the target sound is acoustically identical across both environments, this effect demonstrates the influence of top-down lexical processing in speech perception. Degradations in the auditory signal were hypothesized to render speech stimuli more ambiguous, and therefore promote increased lexical bias. Stimuli included three speech continua that varied by spectral cues of varying speeds, including stop formant transitions (fast), fricative spectra (medium), and vowel formants (slow). Stimuli were presented to listeners with cochlear implants (CIs), and also to listeners with normal hearing with clear spectral quality, or with varying amounts of spectral degradation using a noise vocoder. Results indicated an increased lexical bias effect with degraded speech and for CI listeners, for whom the effect size was related to segment duration. This method can probe an individual's reliance on top-down processing even at the level of simple lexical/phonetic perception.


Assuntos
Implantes Cocleares , Perda Auditiva/fisiopatologia , Fonética , Percepção da Fala , Adulto , Viés , Sinais (Psicologia) , Feminino , Perda Auditiva/reabilitação , Humanos , Masculino , Acústica da Fala
16.
J Acoust Soc Am ; 145(2): 676, 2019 02.
Artigo em Inglês | MEDLINE | ID: mdl-30823808

RESUMO

Accurate perception of binaural cues is essential for left-right sound localization. Much literature focuses on threshold measures of perceptual acuity and accuracy. This study focused on supra-threshold perception using an anticipatory eye movement (AEM) paradigm designed to capture subtle aspects of perception that might not emerge in behavioral-motor responses, such as the accumulation of certainty, and rapid revisions in decision-making. Participants heard interaural timing differences (ITDs) or interaural level differences in correlated or uncorrelated narrowband noises, respectively. A cartoon ball moved behind an occluder and then emerged from the left or right side, consistent with the binaural cue. Participants anticipated the correct answer (before it appeared) by looking where the ball would emerge. Results showed quicker and more steadfast gaze fixations for stimuli with larger cue magnitudes. More difficult stimuli elicited a wider distribution of saccade times and greater number of corrective saccades before final judgment, implying perceptual uncertainty or competition. Cue levels above threshold elicited some wrong-way saccades that were quickly corrected. Saccades to ITDs were earlier and more reliable for low-frequency noises. The AEM paradigm reveals the time course of uncertainty and changes in perceptual decision-making for supra-threshold binaural stimuli even when behavioral responses are consistently correct.


Assuntos
Antecipação Psicológica/fisiologia , Fixação Ocular/fisiologia , Audição/fisiologia , Localização de Som/fisiologia , Adolescente , Adulto , Sinais (Psicologia) , Feminino , Humanos , Masculino , Incerteza , Adulto Jovem
17.
Ear Hear ; 38(1): 42-56, 2017.
Artigo em Inglês | MEDLINE | ID: mdl-27556521

RESUMO

OBJECTIVES: Previous research has found that relative to their peers with normal hearing (NH), children with cochlear implants (CIs) produce the sibilant fricatives /s/ and /∫/ less accurately and with less subphonemic acoustic contrast. The present study sought to further investigate these differences across groups in two ways. First, subphonemic acoustic properties were investigated in terms of dynamic acoustic features that indexed more than just the contrast between /s/ and /∫/. Second, the authors investigated whether such differences in subphonemic acoustic contrast between sibilant fricatives affected the intelligibility of sibilant-initial single word productions by children with CIs and their peers with NH. DESIGN: In experiment 1, productions of /s/ and /∫/ in word-initial prevocalic contexts were elicited from 22 children with bilateral CIs (aged 4 to 7 years) who had at least 2 years of CI experience and from 22 chronological age-matched peers with NH. Acoustic features were measured from 17 points across the fricatives: peak frequency was measured to index the place of articulation contrast; spectral variance and amplitude drop were measured to index the degree of sibilance. These acoustic trajectories were fitted with growth-curve models to analyze time-varying spectral change. In experiment 2, phonemically accurate word productions that were elicited in experiment 1 were embedded within four-talker babble and played to 80 adult listeners with NH. Listeners were asked to repeat the words, and their accuracy rate was used as a measure of the intelligibility of the word productions. Regression analyses were run to test which acoustic properties measured in experiment 1 predicted the intelligibility scores from experiment 2. RESULTS: The peak frequency trajectories indicated that the children with CIs produced less acoustic contrast between /s/ and /∫/. Group differences were observed in terms of the dynamic aspects (i.e., the trajectory shapes) of the acoustic properties. In the productions by children with CIs, the peak frequency and the amplitude drop trajectories were shallower, and the spectral variance trajectories were more asymmetric, exhibiting greater increases in variance (i.e., reduced sibilance) near the fricative-vowel boundary. The listeners' responses to the word productions indicated that when produced by children with CIs, /∫/-initial words were significantly more intelligible than /s/-initial words. However, when produced by children with NH, /s/-initial words and /∫/-initial words were equally intelligible. Intelligibility was partially predicted from the acoustic properties (Cox & Snell pseudo-R > 0.190), and the significant predictors were predominantly dynamic, rather than static, ones. CONCLUSIONS: Productions from children with CIs differed from those produced by age-matched NH controls in terms of their subphonemic acoustic properties. The intelligibility of sibilant-initial single-word productions by children with CIs is sensitive to the place of articulation of the initial consonant (/∫/-initial words were more intelligible than /s/-initial words), but productions by children with NH were equally intelligible across both places of articulation. Therefore, children with CIs still exhibit differential production abilities for sibilant fricatives at an age when their NH peers do not.


Assuntos
Implante Coclear/métodos , Surdez/reabilitação , Acústica da Fala , Inteligibilidade da Fala , Acústica , Criança , Pré-Escolar , Implantes Cocleares , Feminino , Humanos , Masculino , Fonética , Fala
18.
Ear Hear ; 37(6): e377-e390, 2016.
Artigo em Inglês | MEDLINE | ID: mdl-27438871

RESUMO

OBJECTIVES: This study was conducted to measure auditory perception by cochlear implant users in the spectral and temporal domains, using tests of either categorization (using speech-based cues) or discrimination (using conventional psychoacoustic tests). The authors hypothesized that traditional nonlinguistic tests assessing spectral and temporal auditory resolution would correspond to speech-based measures assessing specific aspects of phonetic categorization assumed to depend on spectral and temporal auditory resolution. The authors further hypothesized that speech-based categorization performance would ultimately be a superior predictor of speech recognition performance, because of the fundamental nature of speech recognition as categorization. DESIGN: Nineteen cochlear implant listeners and 10 listeners with normal hearing participated in a suite of tasks that included spectral ripple discrimination, temporal modulation detection, and syllable categorization, which was split into a spectral cue-based task (targeting the /ba/-/da/ contrast) and a timing cue-based task (targeting the /b/-/p/ and /d/-/t/ contrasts). Speech sounds were manipulated to contain specific spectral or temporal modulations (formant transitions or voice onset time, respectively) that could be categorized. Categorization responses were quantified using logistic regression to assess perceptual sensitivity to acoustic phonetic cues. Word recognition testing was also conducted for cochlear implant listeners. RESULTS: Cochlear implant users were generally less successful at utilizing both spectral and temporal cues for categorization compared with listeners with normal hearing. For the cochlear implant listener group, spectral ripple discrimination was significantly correlated with the categorization of formant transitions; both were correlated with better word recognition. Temporal modulation detection using 100- and 10-Hz-modulated noise was not correlated either with the cochlear implant subjects' categorization of voice onset time or with word recognition. Word recognition was correlated more closely with categorization of the controlled speech cues than with performance on the psychophysical discrimination tasks. CONCLUSIONS: When evaluating people with cochlear implants, controlled speech-based stimuli are feasible to use in tests of auditory cue categorization, to complement traditional measures of auditory discrimination. Stimuli based on specific speech cues correspond to counterpart nonlinguistic measures of discrimination, but potentially show better correspondence with speech perception more generally. The ubiquity of the spectral (formant transition) and temporal (voice onset time) stimulus dimensions across languages highlights the potential to use this testing approach even in cases where English is not the native language.


Assuntos
Implante Coclear , Sinais (Psicologia) , Surdez/reabilitação , Percepção da Fala , Adulto , Idoso , Idoso de 80 Anos ou mais , Percepção Auditiva , Implantes Cocleares , Feminino , Humanos , Masculino , Pessoa de Meia-Idade , Ruído , Fonética , Psicoacústica , Análise Espectral , Fala , Fatores de Tempo
19.
J Acoust Soc Am ; 140(6): 4404, 2016 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-28039993

RESUMO

Suboptimal interfaces between cochlear implant (CI) electrodes and auditory neurons result in a loss or distortion of spectral information in specific frequency regions, which likely decreases CI users' speech identification performance. This study exploited speech acoustics to model regions of distorted CI frequency transmission to determine the perceptual consequences of suboptimal electrode-neuron interfaces. Normal hearing adults identified naturally spoken vowels and consonants after spectral information was manipulated through a noiseband vocoder: either (1) low-, middle-, or high-frequency regions of information were removed by zeroing the corresponding channel outputs, or (2) the same regions were distorted by splitting filter outputs to neighboring filters. These conditions simulated the detrimental effects of suboptimal CI electrode-neuron interfaces on spectral transmission. Vowel and consonant confusion patterns were analyzed with sequential information transmission, perceptual distance, and perceptual vowel space analyses. Results indicated that both types of spectral manipulation were equally destructive. Loss or distortion of frequency information produced similar effects on phoneme identification performance and confusion patterns. Consonant error patterns were consistently based on place of articulation. Vowel confusions showed that perceptions gravitated away from the degraded frequency region in a predictable manner, indicating that vowels can probe frequency-specific regions of spectral degradations.


Assuntos
Implantes Cocleares , Adulto , Cóclea , Implante Coclear , Feminino , Humanos , Masculino , Neurônios , Fonética , Percepção da Fala , Adulto Jovem
20.
J Acoust Soc Am ; 139(4): 1724, 2016 04.
Artigo em Inglês | MEDLINE | ID: mdl-27106319

RESUMO

Children who use bilateral cochlear implants (BiCIs) show significantly poorer sound localization skills than their normal hearing (NH) peers. This difference has been attributed, in part, to the fact that cochlear implants (CIs) do not faithfully transmit interaural time differences (ITDs) and interaural level differences (ILDs), which are known to be important cues for sound localization. Interestingly, little is known about binaural sensitivity in NH children, in particular, with stimuli that constrain acoustic cues in a manner representative of CI processing. In order to better understand and evaluate binaural hearing in children with BiCIs, the authors first undertook a study on binaural sensitivity in NH children ages 8-10, and in adults. Experiments evaluated sound discrimination and lateralization using ITD and ILD cues, for stimuli with robust envelope cues, but poor representation of temporal fine structure. Stimuli were spondaic words, Gaussian-enveloped tone pulse trains (100 pulse-per-second), and transposed tones. Results showed that discrimination thresholds in children were adult-like (15-389 µs for ITDs and 0.5-6.0 dB for ILDs). However, lateralization based on the same binaural cues showed higher variability than seen in adults. Results are discussed in the context of factors that may be responsible for poor representation of binaural cues in bilaterally implanted children.


Assuntos
Sinais (Psicologia) , Audição , Percepção da Altura Sonora , Localização de Som , Estimulação Acústica/métodos , Acústica , Adulto , Fatores Etários , Audiometria/métodos , Limiar Auditivo , Criança , Implante Coclear/instrumentação , Implantes Cocleares , Discriminação Psicológica , Feminino , Lateralidade Funcional , Humanos , Masculino , Pessoas com Deficiência Auditiva/psicologia , Pessoas com Deficiência Auditiva/reabilitação , Psicoacústica , Espectrografia do Som , Fatores de Tempo
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA