Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 8.394
Filtrar
1.
J Speech Lang Hear Res ; 65(9): 3420-3437, 2022 Sep 12.
Artículo en Inglés | MEDLINE | ID: mdl-36054879

RESUMEN

PURPOSE: This study examines communicative congruence and communicative dysphoria in adults who deny having a voice disorder. Communicative congruence is the degree to which a person's communication is consistent with their sense of self/identity. Communicative dysphoria is the psychological entropy resulting from communicative incongruence. We propose that these experiences may influence patients' psychosocial well-being and are thus relevant to the field of speech-language pathology. We hypothesized that both constructs would be normally distributed with an inverse relationship. We also hypothesized that communicative congruence would predict scores on the Center for Epidemiological Studies-Depression (CES-D) scale, subscales of the Big Five Aspect Scales (BFAS; a personality measure), and achieve convergent validity with the Vocal Congruence Scale (VCS). METHOD: Participants (adults 18-70 years) completed the BFAS, CES-D, VCS, and the Voice Handicap Index-10 (VHI-10) before recording a series of speech tasks. Participants' recordings were played back while they responded to questions probing their communicative congruence and communicative dysphoria. RESULTS: The 196 participants were predominantly female (67.3%) and cisgender (96.4%). Communicative congruence was negatively skewed, and communicative dysphoria was normally distributed. Both variables significantly related to each other: More incongruence was associated with more dysphoria. Communicative congruence was inversely related to CES-D scores. The personality metatrait Plasticity related to communicative congruence, as did the domain of Extraversion and the aspects Withdrawal, Enthusiasm, and Assertiveness. Communicative congruence achieved high convergent validity with the VCS. CONCLUSIONS: Counter to our hypothesis, participants reported more congruence than incongruence, but all other hypotheses were supported. Participants who reported incongruence were more likely to report symptoms of depression. These findings suggest that even in a predominantly cisgender cohort, some individuals' mental well-being might relate to how well their communication aligns with their identity. This work may inform future investigations into these constructs and their effects on voice therapy outcomes. SUPPLEMENTAL MATERIAL: https://doi.org/10.23641/asha.20739967.


Asunto(s)
Patología del Habla y Lenguaje , Trastornos de la Voz , Voz , Adulto , Comunicación , Femenino , Humanos , Masculino , Habla , Trastornos de la Voz/diagnóstico , Trastornos de la Voz/psicología
2.
Annu Int Conf IEEE Eng Med Biol Soc ; 2022: 998-1001, 2022 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-36086187

RESUMEN

This work focuses on the automatic detection of COVID-19 from the analysis of vocal sounds, including sustained vowels, coughs, and speech while reading a short text. Specifically, we use the Mel-spectrogram representations of these acoustic signals to train neural network-based models for the task at hand. The extraction of deep learnt representations from the Mel-spectrograms is performed with Convolutional Neural Networks (CNNs). In an attempt to guide the training of the embedded representations towards more separable and robust inter-class representations, we explore the use of a triplet loss function. The experiments performed are conducted using the Your Voice Counts dataset, a new dataset containing German speakers collected using smartphones. The results obtained support the suitability of using triplet loss-based models to detect COVID-19 from vocal sounds. The best Unweighted Average Recall (UAR) of 66.5 % is obtained using a triplet loss-based model exploiting vocal sounds recorded while reading.


Asunto(s)
COVID-19 , Voz , Acústica , COVID-19/diagnóstico , Humanos , Redes Neurales de la Computación , Habla
3.
J Acoust Soc Am ; 152(2): 820, 2022 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-36050169

RESUMEN

Different speakers produce the same intended vowel with very different physical properties. Fundamental frequency (F0) and formant frequencies (FF), the two main parameters that discriminate between voices, also influence vowel perception. While it has been shown that listeners comprehend speech more accurately if they are familiar with a talker's voice, it is still unclear how such prior information is used when decoding the speech stream. In three online experiments, we examined the influence of speaker context via F0 and FF shifts on the perception of /o/-/u/ vowel contrasts. Participants perceived vowels from an /o/-/u/ continuum shifted toward /u/ when F0 was lowered or FF increased relative to the original speaker's voice and vice versa. This shift was reduced when the speakers were presented in a block-wise context compared to random order. Conversely, the original base voice was perceived to be shifted toward /u/ when presented in the context of a low F0 or high FF speaker, compared to a shift toward /o/ with high F0 or low FF speaker context. These findings demonstrate that that F0 and FF jointly influence vowel perception in speaker context.


Asunto(s)
Percepción del Habla , Voz , Humanos , Fonética , Habla , Acústica del Lenguaje
4.
Sensors (Basel) ; 22(17)2022 Sep 02.
Artículo en Inglés | MEDLINE | ID: mdl-36081092

RESUMEN

Deep learning techniques such as convolutional neural networks (CNN) have been successfully applied to identify pathological voices. However, the major disadvantage of using these advanced models is the lack of interpretability in explaining the predicted outcomes. This drawback further introduces a bottleneck for promoting the classification or detection of voice-disorder systems, especially in this pandemic period. In this paper, we proposed using a series of learnable sinc functions to replace the very first layer of a commonly used CNN to develop an explainable SincNet system for classifying or detecting pathological voices. The applied sinc filters, a front-end signal processor in SincNet, are critical for constructing the meaningful layer and are directly used to extract the acoustic features for following networks to generate high-level voice information. We conducted our tests on three different Far Eastern Memorial Hospital voice datasets. From our evaluations, the proposed approach achieves the highest 7%-accuracy and 9%-sensitivity improvements from conventional methods and thus demonstrates superior performance in predicting input pathological waveforms of the SincNet system. More importantly, we intended to give possible explanations between the system output and the first-layer extracted speech features based on our evaluated results.


Asunto(s)
Trastornos de la Voz , Voz , Acústica , Humanos , Redes Neurales de la Computación , Trastornos de la Voz/diagnóstico
5.
Sensors (Basel) ; 22(18)2022 Sep 13.
Artículo en Inglés | MEDLINE | ID: mdl-36146251

RESUMEN

Monitoring cognitive workload has the potential to improve both the performance and fidelity of human decision making. However, previous efforts towards discriminating further than binary levels (e.g., low/high or neutral/high) in cognitive workload classification have not been successful. This lack of sensitivity in cognitive workload measurements might be due to individual differences as well as inadequate methodology used to analyse the measured signal. In this paper, a method that combines the speech signal with cardiovascular measurements for screen and heartbeat classification is introduced. For validation, speech and cardiovascular signals from 97 university participants and 20 airline pilot participants were collected while cognitive stimuli of varying difficulty level were induced with the Stroop colour/word test. For the trinary classification scheme (low, medium, high cognitive workload) the prominent result using classifiers trained on each participant achieved 15.17 ± 0.79% and 17.38 ± 1.85% average misclassification rates indicating good discrimination at three levels of cognitive workload. Combining cardiovascular and speech measures synchronized to each heartbeat and consolidated with short-term dynamic measures might therefore provide enhanced sensitivity in cognitive workload monitoring. The results show that the influence of individual differences is a limiting factor for a generic classification and highlights the need for research to focus on methods that incorporate individual differences to achieve even better results. This method can potentially be used to measure and monitor workload in real time in operational environments.


Asunto(s)
Voz , Carga de Trabajo , Cognición , Frecuencia Cardíaca , Humanos , Habla , Carga de Trabajo/psicología
6.
Artículo en Inglés | MEDLINE | ID: mdl-36141675

RESUMEN

In general, it is common knowledge that people's feelings are reflected in their voice and facial expressions. This research work focuses on developing techniques for diagnosing depression based on acoustic properties of the voice. In this study, we developed a composite index of vocal acoustic properties that can be used for depression detection. Voice recordings were collected from patients undergoing outpatient treatment for major depressive disorder at a hospital or clinic following a physician's diagnosis. Numerous features were extracted from the collected audio data using openSMILE software. Furthermore, qualitatively similar features were combined using principal component analysis. The resulting components were incorporated as parameters in a logistic regression based classifier, which achieved a diagnostic accuracy of ~90% on the training set and ~80% on the test set. Lastly, the proposed metric could serve as a new measure for evaluation of major depressive disorder.


Asunto(s)
Trastorno Depresivo Mayor , Trastornos de la Voz , Voz , Acústica , Trastorno Depresivo Mayor/diagnóstico , Humanos , Modelos Logísticos
7.
PLoS One ; 17(9): e0266991, 2022.
Artículo en Inglés | MEDLINE | ID: mdl-36156597

RESUMEN

INTRODUCTION: Group musical activities using wind instruments have been restricted during the CoVID19 pandemic due to suspected higher risk of virus transmission. It was presumed that the aerosols exhaled through the tubes while playing would be ejected over larger distances and spread into the room due to jet stream effects. In particular, the soprano recorder is widely used as an instrument in school classes, for beginners of all age groups in their musical education, in the context of leisure activities and in professional concert performances. Understanding the aerosol impulse dispersion characteristics of playing the soprano recorder could assist with the establishment of concepts for safe music-making. METHODS: Five adult professionally trained soprano recorder players (4 female, 1 male) played four bars of the main theme of L. van Beethoven's "Ode to Joy" in low and in high octaves, as well as with 3 different potential protection devices in the high octave. For comparison they spoke the corresponding text by F. Schiller. Before each task, they inhaled .5 L of vapor from an e-cigarette filled with base liquid. The vapor cloud escaping during speaking or playing was recorded by cameras and its spread was measured as a function of time in the three spatial dimensions. The potential safety devices were rated for practicability with a questionnaire, and their influence on the sound was compared, generating a long-term average spectrum from the audio data. RESULTS: When playing in the high octave, at the end of the task the clouds showed a median distance of 1.06 m to the front and .57 m diameter laterally (maxima: x: 1.35 m and y: .97 m). It was found that the clouds' expansion values in playing the recorder with and without safety measures are mostly lower when compared to the ordinary, raised speaking voice of the same subjects. The safety devices which covered the instrument did not show clear advantages and were rated as unpractical by the subjects. The most effective reduction of the cloud was reached when playing into a suction funnel. CONCLUSION: The aerosol dispersion characteristics of soprano recorders seem comparable to clarinets. The tested safety devices which covered holes of the instrument did not show clear benefits.


Asunto(s)
COVID-19 , Sistemas Electrónicos de Liberación de Nicotina , Música , Voz , Adulto , COVID-19/epidemiología , COVID-19/prevención & control , Femenino , Humanos , Masculino , Aerosoles y Gotitas Respiratorias
8.
JASA Express Lett ; 2(4): 045204, 2022 04.
Artículo en Inglés | MEDLINE | ID: mdl-36154231

RESUMEN

This study examined how speaking style and guise influence the intelligibility of text-to-speech (TTS) and naturally produced human voices. Results showed that TTS voices were less intelligible overall. Although using a clear speech style improved intelligibility for both human and TTS voices (using "newscaster" neural TTS), the clear speech effect was stronger for TTS voices. Finally, a visual device guise decreased intelligibility, regardless of voice type. The results suggest that both speaking style and visual guise affect intelligibility of human and TTS voices. Findings are discussed in terms of theories about the role of social information in speech perception.


Asunto(s)
Percepción del Habla , Envío de Mensajes de Texto , Voz , Cognición , Humanos , Inteligibilidad del Habla
9.
JASA Express Lett ; 2(3): 035203, 2022 03.
Artículo en Inglés | MEDLINE | ID: mdl-36154631

RESUMEN

While speaking, hand postures, such as holding a hand in front of the mouth or cupping the hands around the mouth, influence human voice directivity. This study presents and analyzes spherical voice directivity datasets of an articulated [a] with and without hand postures. The datasets were determined from measurements with 13 subjects in a surrounding spherical microphone array with 32 microphones and then upsampled to a higher spatial resolution. The results show that hand postures strongly impact voice directivity and affect the directivity index by up to 6 dB, which is more than variances caused by phoneme-dependent differences.


Asunto(s)
Voz , Humanos , Postura
10.
PLoS One ; 17(9): e0274595, 2022.
Artículo en Inglés | MEDLINE | ID: mdl-36103566

RESUMEN

This investigation tested the effect of priming on pareidolia (the hearing of illusory words in ambiguous stimuli). Participants (41 women, 20 men, mean age 29.95 years) were assigned to primed (n = 30) or unprimed (n = 31) groups: the former were told the study was of 'purported ghosts voices', the latter 'voices in noisy environments.' Participants were assessed for perception of human voices within recordings of purported electronic voice phenomena (EVP), degraded human speech, normal human speech, and white noise. The primed group had significantly higher perception of voices within EVPs than in degraded speech, this difference was not found for unprimed participants. In contrast to the previous use of this design, the primed group did not have higher perception of voices in EVPs and degraded speech than did the unprimed group. The Aesthetic Sensitivity dimension of the Highly Sensitive Person Scale (HSPS) was associated with detection of degraded stimuli, but not with accuracy of stimulus identification. HSPS score was related to lifetime reporting of anomalous and paranormal experiences. This study partially replicates a paranormal priming effect and shows relationships between HSPS and detection of ambiguous stimuli and anomalous and paranormal experiences.


Asunto(s)
Ilusiones , Voz , Adulto , Femenino , Audición , Humanos , Masculino , Habla
11.
Philos Trans R Soc Lond B Biol Sci ; 377(1863): 20210511, 2022 Nov 07.
Artículo en Inglés | MEDLINE | ID: mdl-36126659

RESUMEN

A substantial body of acoustic and behavioural evidence points to the existence of two broad categories of laughter in humans: spontaneous laughter that is emotionally genuine and somewhat involuntary, and volitional laughter that is produced on demand. In this study, we tested the hypothesis that these are also physiologically distinct vocalizations, by measuring and comparing them using real-time magnetic resonance imaging (rtMRI) of the vocal tract. Following Ruch and Ekman (Ruch and Ekman 2001 In Emotions, qualia, and consciousness (ed. A Kaszniak), pp. 426-443), we further predicted that spontaneous laughter should be relatively less speech-like (i.e. less articulate) than volitional laughter. We collected rtMRI data from five adult human participants during spontaneous laughter, volitional laughter and spoken vowels. We report distinguishable vocal tract shapes during the vocalic portions of these three vocalization types, where volitional laughs were intermediate between spontaneous laughs and vowels. Inspection of local features within the vocal tract across the different vocalization types offers some additional support for Ruch and Ekman's predictions. We discuss our findings in light of a dual pathway hypothesis for the neural control of human volitional and spontaneous vocal behaviours, identifying tongue shape and velum lowering as potential biomarkers of spontaneous laughter to be investigated in future research. This article is part of the theme issue 'Cracking the laugh code: laughter through the lens of biology, psychology and neuroscience'.


Asunto(s)
Risa , Voz , Adulto , Emociones , Humanos , Risa/fisiología , Risa/psicología , Imagen por Resonancia Magnética , Volición
12.
Philos Trans R Soc Lond B Biol Sci ; 377(1863): 20210179, 2022 Nov 07.
Artículo en Inglés | MEDLINE | ID: mdl-36126665

RESUMEN

Like most human non-verbal vocalizations, laughter is produced by speakers of all languages, across all known societies. But despite this obvious fact (or perhaps because of it), there is little comparative research examining the structural and functional similarity of laughter across speakers from different cultures. Here, we describe existing research examining (i) the perception of laughter across disparate cultures, (ii) conversation analysis examining how laughter manifests itself during discourse across different languages, and (iii) computational methods developed for automatically detecting laughter in spoken language databases. Together, these three areas of investigation provide clues regarding universals and cultural variations in laughter production and perception, and offer methodological tools that can be useful for future large-scale cross-cultural studies. We conclude by providing suggestions for areas of research and predictions of what we should expect to discover. Overall, we highlight how important questions regarding human vocal communication across cultures can be addressed through the examination of spontaneous and volitional laughter. This article is part of the theme issue 'Cracking the laugh code: laughter through the lens of biology, psychology and neuroscience'.


Asunto(s)
Risa , Voz , Comunicación , Comparación Transcultural , Humanos , Lenguaje , Risa/psicología
13.
Sensors (Basel) ; 22(16)2022 Aug 12.
Artículo en Inglés | MEDLINE | ID: mdl-36015789

RESUMEN

Pitch estimation is widely used in speech and audio signal processing. However, the current methods of modeling harmonic structure used for pitch estimation cannot always match the harmonic distribution of actual signals. Due to the structure of vocal tract, the acoustic nature of musical equipment, and the spectrum leakage issue, speech and audio signals' harmonic frequencies often slightly deviate from the integer multiple of the pitch. This paper starts with the summation of residual harmonics (SRH) method and makes two main modifications. First, the spectral peak position constraint of strict integer multiple is modified to allow slight deviation, which benefits capturing harmonics. Second, a main pitch segment extension scheme with low computational cost feature is proposed to utilize the smooth prior of pitch more efficiently. Besides, the pitch segment extension scheme is also integrated into the SRH method's voiced/unvoiced decision to reduce short-term errors. Accuracy comparison experiments with ten pitch estimation methods show that the proposed method has better overall accuracy and robustness. Time cost experiments show that the time cost of the proposed method reduces to around 1/8 of the state-of-the-art fast NLS method on the experimental computer.


Asunto(s)
Voz , Computadores , Procesamiento de Señales Asistido por Computador
14.
Sensors (Basel) ; 22(16)2022 Aug 13.
Artículo en Inglés | MEDLINE | ID: mdl-36015817

RESUMEN

Silent speech interfaces (SSIs) convert non-audio bio-signals, such as articulatory movement, to speech. This technology has the potential to recover the speech ability of individuals who have lost their voice but can still articulate (e.g., laryngectomees). Articulation-to-speech (ATS) synthesis is an algorithm design of SSI that has the advantages of easy-implementation and low-latency, and therefore is becoming more popular. Current ATS studies focus on speaker-dependent (SD) models to avoid large variations of articulatory patterns and acoustic features across speakers. However, these designs are limited by the small data size from individual speakers. Speaker adaptation designs that include multiple speakers' data have the potential to address the issue of limited data size from single speakers; however, few prior studies have investigated their performance in ATS. In this paper, we investigated speaker adaptation on both the input articulation and the output acoustic signals (with or without direct inclusion of data from test speakers) using the publicly available electromagnetic articulatory (EMA) dataset. We used Procrustes matching and voice conversion for articulation and voice adaptation, respectively. The performance of the ATS models was measured objectively by the mel-cepstral distortions (MCDs). The synthetic speech samples were generated and are provided in the supplementary material. The results demonstrated the improvement brought by both Procrustes matching and voice conversion on speaker-independent ATS. With the direct inclusion of target speaker data in the training process, the speaker-adaptive ATS achieved a comparable performance to speaker-dependent ATS. To our knowledge, this is the first study that has demonstrated that speaker-adaptive ATS can achieve a non-statistically different performance to speaker-dependent ATS.


Asunto(s)
Percepción del Habla , Voz , Acústica , Humanos , Habla , Acústica del Lenguaje
15.
Clin Linguist Phon ; 36(7): 630-641, 2022 Jul 03.
Artículo en Inglés | MEDLINE | ID: mdl-36005746

RESUMEN

This study presents a nonlinear phonological analysis of speech data from a Farsi-speaking child with protracted phonological development (aged 4;8) with very low accuracy on consonants. Results revealed some common phenomena (fricatives produced as stops; dorsals and non-anterior coronals produced as anterior coronals) and some uncommon phenomena (nasals produced as oral stops; voicing and devoicing of singleton obstruents in all word positions). These phenomena interacted in word-medial clusters to create an unusual sequence of two anterior-coronal or two bilabial stops, with C1 voiced and C2 voiceless, clusters which do not occur in the basic phonology of Farsi spoken by adults. We present a non-linear constraints-based analysis of the child's speech production and a plan for intervention targeting the child's special difficulties.


Asunto(s)
Fonética , Voz , Adulto , Niño , Humanos , Habla , Medición de la Producción del Habla
16.
Emotion ; 22(6): 1394-1399, 2022 Sep.
Artículo en Inglés | MEDLINE | ID: mdl-36006704

RESUMEN

A basic premise of classic emotion theories is that distinct emotional experiences yield distinct emotional vocalizations-each informative of its situational context. Furthermore, it is commonly assumed that emotional vocalizations become more distinct and diagnostic as their intensity increases. Critically, these theoretical assumptions largely rely on research utilizing posed vocal reactions of actors, which may be overly simplified and stereotypical. While recent work suggests that intense, real-life vocalizations may be nondiagnostic, the exact way in which increasing degrees of situational intensity affect the perceived valence of real-life versus posed expressions remains unknown. Here we compared real-life and posed vocalizations to winning increasing amounts of money in the lottery. Results show that while posed vocalizations are perceived as positive for both low- and high-sum wins, real-life vocalizations are perceived as positive only for low-sum wins, but as negative for high-sum wins. These findings demonstrate the potential gaps between real-life and posed expressions and highlight the role of situational intensity in driving perceptual ambiguity for real-life emotional expressions. (PsycInfo Database Record (c) 2022 APA, all rights reserved).


Asunto(s)
Emociones , Voz , Humanos
18.
J Acoust Soc Am ; 152(1): 302, 2022 07.
Artículo en Inglés | MEDLINE | ID: mdl-35931496

RESUMEN

Voice therapy often emphasizes vibratory sensations in the front part of the vocal tract during phonation to improve vocal efficiency. It remains unclear what laryngeal and vocal tract adjustments are elicited in speakers by this emphasis on oral vibratory sensations. Using a three-dimensional phonation model, this study aims to identify laryngeal and epilaryngeal adjustments that might produce maximal oral vibratory sensations during phonation, as quantified by the oral sound pressure level (SPL), and thus are likely to be elicited in voice therapy at different semi-occluded vocal tract configurations. Results show that maximum oral SPL occurs at intermediate vocal fold adduction configurations characterized by a trade-off between glottal gap and vocal fold vertical thickness. Epilaryngeal tube narrowing further increases the oral SPL in an open vocal tract, but has little effect on oral SPL in semi-occluded vocal tracts. Laryngeal and epilaryngeal configurations producing the maximum oral SPL generally have lower peak vocal fold contact pressure when producing a target output SPL. These favorable configurations are more easily identified in open vocal tracts than semi-occluded vocal tracts. However, semi-occlusion increases both the mean and dynamic oral pressure, which may familiarize speakers with oral vibratory sensations and facilitate adoption of favorable laryngeal configurations.


Asunto(s)
Laringe , Voz , Humanos , Fonación , Sensación , Vibración , Pliegues Vocales
19.
J Speech Lang Hear Res ; 65(8): 3165-3194, 2022 08 17.
Artículo en Inglés | MEDLINE | ID: mdl-35926089

RESUMEN

OBJECTIVE: Some cochlear implant (CI) users report having difficulty accessing indexical information in the speech signal, presumably due to limitations in the transmission of fine spectrotemporal cues. The purpose of this review article was to systematically review and evaluate the existing research on talker processing in CI users. Specifically, we reviewed the performance of CI users in three types of talker- and voice-related tasks. We also examined the different factors (such as participant, hearing, and device characteristics) that might influence performance in these specific tasks. DESIGN: We completed a systematic search of the literature with select key words using citation aggregation software to search Google Scholar. We included primary reports that tested (a) talker discrimination, (b) voice perception, and (c) talker identification. Each report must have had at least one group of participants with CIs. Each included study was also evaluated for quality of evidence. RESULTS: The searches resulted in 1,561 references, which were first screened for inclusion and then evaluated in full. Forty-three studies examining talker discrimination, voice perception, and talker identification were included in the final review. Most studies were focused on postlingually deafened and implanted adult CI users, with fewer studies focused on prelingual implant users. In general, CI users performed above chance in these tasks. When there was a difference between groups, CI users performed less accurately than their normal-hearing (NH) peers. A subset of CI users reached the same level of performance as NH participants exposed to noise-vocoded stimuli. Some studies found that CI users and NH participants relied on different cues for talker perception. Within groups of CI users, there is moderate evidence for a bimodal benefit for talker processing, and there are mixed findings about the effects of hearing experience. CONCLUSIONS: The current review highlights the challenges faced by CI users in tracking and recognizing voices and how they adapt to it. Although large variability exists, there is evidence that CI users can process indexical information from speech, though with less accuracy than their NH peers. Recent work has described some of the factors that might ease the challenges of talker processing in CI users. We conclude by suggesting some future avenues of research to optimize real-world speech outcomes.


Asunto(s)
Implantación Coclear , Implantes Cocleares , Sordera , Percepción del Habla , Voz , Adulto , Sordera/rehabilitación , Sordera/cirugía , Humanos
20.
J Speech Lang Hear Res ; 65(8): 2881-2895, 2022 08 17.
Artículo en Inglés | MEDLINE | ID: mdl-35930680

RESUMEN

PURPOSE: This exploratory study aims to investigate variations in voice production in the presence of background noise (Lombard effect) in individuals with nonphonotraumatic vocal hyperfunction (NPVH) and individuals with typical voices using acoustic, aerodynamic, and vocal fold vibratory measures of phonatory function. METHOD: Nineteen participants with NPVH and 19 participants with typical voices produced simple vocal tasks in three sequential background conditions: baseline (in quiet), Lombard (in noise), and recovery (5 min after removing the noise). The Lombard condition consisted of speech-shaped noise at 80 dB SPL through audiometric headphones. Acoustic measures from a microphone, glottal aerodynamic parameters estimated from the oral airflow measured with a circumferentially vented pneumotachograph mask, and vocal fold vibratory parameters from high-speed videoendoscopy were analyzed. RESULTS: During the Lombard condition, both groups exhibited a decrease in open quotient and increases in sound pressure level, peak-to-peak glottal airflow, maximum flow declination rate, and subglottal pressure. During the recovery condition, the acoustic and aerodynamic measures of individuals with typical voices returned to those of the baseline condition; however, recovery measures for individuals with NPVH did not return to baseline values. CONCLUSIONS: As expected, individuals with NPVH and participants with typical voices exhibited a Lombard effect in the presence of elevated background noise levels. During the recovery condition, individuals with NPVH did not return to their baseline state, pointing to a persistence of the Lombard effect after noise removal. This behavior could be related to disruptions in laryngeal motor control and may play a role in the etiology of NPVH. SUPPLEMENTAL MATERIAL: https://doi.org/10.23641/asha.20415600.


Asunto(s)
Pliegues Vocales , Voz , Acústica , Glotis , Humanos , Fonación
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA
...