Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Resultados 1 - 20 de 1.353
Filtrar
Más filtros

Publication year range
1.
Proc Natl Acad Sci U S A ; 121(26): e2318361121, 2024 Jun 25.
Artículo en Inglés | MEDLINE | ID: mdl-38889147

RESUMEN

When listeners hear a voice, they rapidly form a complex first impression of who the person behind that voice might be. We characterize how these multivariate first impressions from voices emerge over time across different levels of abstraction using electroencephalography and representational similarity analysis. We find that for eight perceived physical (gender, age, and health), trait (attractiveness, dominance, and trustworthiness), and social characteristics (educatedness and professionalism), representations emerge early (~80 ms after stimulus onset), with voice acoustics contributing to those representations between ~100 ms and 400 ms. While impressions of person characteristics are highly correlated, we can find evidence for highly abstracted, independent representations of individual person characteristics. These abstracted representationse merge gradually over time. That is, representations of physical characteristics (age, gender) arise early (from ~120 ms), while representations of some trait and social characteristics emerge later (~360 ms onward). The findings align with recent theoretical models and shed light on the computations underpinning person perception from voices.


Asunto(s)
Percepción Auditiva , Encéfalo , Electroencefalografía , Voz , Humanos , Masculino , Femenino , Voz/fisiología , Adulto , Encéfalo/fisiología , Percepción Auditiva/fisiología , Adulto Joven , Percepción Social
2.
Proc Natl Acad Sci U S A ; 121(25): e2405588121, 2024 Jun 18.
Artículo en Inglés | MEDLINE | ID: mdl-38861607

RESUMEN

Many animals can extract useful information from the vocalizations of other species. Neuroimaging studies have evidenced areas sensitive to conspecific vocalizations in the cerebral cortex of primates, but how these areas process heterospecific vocalizations remains unclear. Using fMRI-guided electrophysiology, we recorded the spiking activity of individual neurons in the anterior temporal voice patches of two macaques while they listened to complex sounds including vocalizations from several species. In addition to cells selective for conspecific macaque vocalizations, we identified an unsuspected subpopulation of neurons with strong selectivity for human voice, not merely explained by spectral or temporal structure of the sounds. The auditory representational geometry implemented by these neurons was strongly related to that measured in the human voice areas with neuroimaging and only weakly to low-level acoustical structure. These findings provide new insights into the neural mechanisms involved in auditory expertise and the evolution of communication systems in primates.


Asunto(s)
Percepción Auditiva , Imagen por Resonancia Magnética , Neuronas , Vocalización Animal , Voz , Animales , Humanos , Neuronas/fisiología , Voz/fisiología , Imagen por Resonancia Magnética/métodos , Vocalización Animal/fisiología , Percepción Auditiva/fisiología , Masculino , Macaca mulatta , Encéfalo/fisiología , Estimulación Acústica , Mapeo Encefálico/métodos
3.
PLoS Comput Biol ; 20(2): e1011849, 2024 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-38315733

RESUMEN

Sleep deprivation has an ever-increasing impact on individuals and societies. Yet, to date, there is no quick and objective test for sleep deprivation. Here, we used automated acoustic analyses of the voice to detect sleep deprivation. Building on current machine-learning approaches, we focused on interpretability by introducing two novel ideas: the use of a fully generic auditory representation as input feature space, combined with an interpretation technique based on reverse correlation. The auditory representation consisted of a spectro-temporal modulation analysis derived from neurophysiology. The interpretation method aimed to reveal the regions of the auditory representation that supported the classifiers' decisions. Results showed that generic auditory features could be used to detect sleep deprivation successfully, with an accuracy comparable to state-of-the-art speech features. Furthermore, the interpretation revealed two distinct effects of sleep deprivation on the voice: changes in slow temporal modulations related to prosody and changes in spectral features related to voice quality. Importantly, the relative balance of the two effects varied widely across individuals, even though the amount of sleep deprivation was controlled, thus confirming the need to characterize sleep deprivation at the individual level. Moreover, while the prosody factor correlated with subjective sleepiness reports, the voice quality factor did not, consistent with the presence of both explicit and implicit consequences of sleep deprivation. Overall, the findings show that individual effects of sleep deprivation may be observed in vocal biomarkers. Future investigations correlating such markers with objective physiological measures of sleep deprivation could enable "sleep stethoscopes" for the cost-effective diagnosis of the individual effects of sleep deprivation.


Asunto(s)
Privación de Sueño , Voz , Humanos , Sueño , Calidad de la Voz , Vigilia
4.
Cereb Cortex ; 34(1)2024 01 14.
Artículo en Inglés | MEDLINE | ID: mdl-38142293

RESUMEN

Selective attention to one speaker in multi-talker environments can be affected by the acoustic and semantic properties of speech. One highly ecological feature of speech that has the potential to assist in selective attention is voice familiarity. Here, we tested how voice familiarity interacts with selective attention by measuring the neural speech-tracking response to both target and non-target speech in a dichotic listening "Cocktail Party" paradigm. We measured Magnetoencephalography from n = 33 participants, presented with concurrent narratives in two different voices, and instructed to pay attention to one ear ("target") and ignore the other ("non-target"). Participants were familiarized with one of the voices during the week prior to the experiment, rendering this voice familiar to them. Using multivariate speech-tracking analysis we estimated the neural responses to both stimuli and replicate their well-established modulation by selective attention. Importantly, speech-tracking was also affected by voice familiarity, showing enhanced response for target speech and reduced response for non-target speech in the contra-lateral hemisphere, when these were in a familiar vs. an unfamiliar voice. These findings offer valuable insight into how voice familiarity, and by extension, auditory-semantics, interact with goal-driven attention, and facilitate perceptual organization and speech processing in noisy environments.


Asunto(s)
Percepción del Habla , Voz , Humanos , Habla , Percepción del Habla/fisiología , Reconocimiento en Psicología/fisiología , Semántica
5.
Eur J Neurosci ; 60(2): 4078-4094, 2024 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-38777332

RESUMEN

Although the attractiveness of voices plays an important role in social interactions, it is unclear how voice attractiveness and social interest influence social decision-making. Here, we combined the ultimatum game with recording event-related brain potentials (ERPs) and examined the effect of attractive versus unattractive voices of the proposers, expressing positive versus negative social interest ("I like you" vs. "I don't like you"), on the acceptance of the proposal. Overall, fair offers were accepted at significantly higher rates than unfair offers, and high voice attractiveness increased acceptance rates for all proposals. In ERPs in response to the voices, their attractiveness and expressed social interests yielded early additive effects in the N1 component, followed by interactions in the subsequent P2, P3 and N400 components. More importantly, unfair offers elicited a larger Medial Frontal Negativity (MFN) than fair offers but only when the proposer's voice was unattractive or when the voice carried positive social interest. These results suggest that both voice attractiveness and social interest moderate social decision-making and there is a similar "beauty premium" for voices as for faces.


Asunto(s)
Toma de Decisiones , Potenciales Evocados , Voz , Humanos , Masculino , Femenino , Potenciales Evocados/fisiología , Voz/fisiología , Toma de Decisiones/fisiología , Adulto Joven , Adulto , Electroencefalografía/métodos , Encéfalo/fisiología , Adolescente
6.
Hum Brain Mapp ; 45(10): e26724, 2024 Jul 15.
Artículo en Inglés | MEDLINE | ID: mdl-39001584

RESUMEN

Music is ubiquitous, both in its instrumental and vocal forms. While speech perception at birth has been at the core of an extensive corpus of research, the origins of the ability to discriminate instrumental or vocal melodies is still not well investigated. In previous studies comparing vocal and musical perception, the vocal stimuli were mainly related to speaking, including language, and not to the non-language singing voice. In the present study, to better compare a melodic instrumental line with the voice, we used singing as a comparison stimulus, to reduce the dissimilarities between the two stimuli as much as possible, separating language perception from vocal musical perception. In the present study, 45 newborns were scanned, 10 full-term born infants and 35 preterm infants at term-equivalent age (mean gestational age at test = 40.17 weeks, SD = 0.44) using functional magnetic resonance imaging while listening to five melodies played by a musical instrument (flute) or sung by a female voice. To examine the dynamic task-based effective connectivity, we employed a psychophysiological interaction of co-activation patterns (PPI-CAPs) analysis, using the auditory cortices as seed region, to investigate moment-to-moment changes in task-driven modulation of cortical activity during an fMRI task. Our findings reveal condition-specific, dynamically occurring patterns of co-activation (PPI-CAPs). During the vocal condition, the auditory cortex co-activates with the sensorimotor and salience networks, while during the instrumental condition, it co-activates with the visual cortex and the superior frontal cortex. Our results show that the vocal stimulus elicits sensorimotor aspects of the auditory perception and is processed as a more salient stimulus while the instrumental condition activated higher-order cognitive and visuo-spatial networks. Common neural signatures for both auditory stimuli were found in the precuneus and posterior cingulate gyrus. Finally, this study adds knowledge on the dynamic brain connectivity underlying the newborns capability of early and specialized auditory processing, highlighting the relevance of dynamic approaches to study brain function in newborn populations.


Asunto(s)
Percepción Auditiva , Imagen por Resonancia Magnética , Música , Humanos , Femenino , Masculino , Percepción Auditiva/fisiología , Recién Nacido , Canto/fisiología , Recien Nacido Prematuro/fisiología , Mapeo Encefálico , Estimulación Acústica , Encéfalo/fisiología , Encéfalo/diagnóstico por imagen , Voz/fisiología
7.
Psychol Sci ; 35(3): 250-262, 2024 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-38289294

RESUMEN

Fundamental frequency ( fo) is the most perceptually salient vocal acoustic parameter, yet little is known about how its perceptual influence varies across societies. We examined how fo affects key social perceptions and how socioecological variables modulate these effects in 2,647 adult listeners sampled from 44 locations across 22 nations. Low male fo increased men's perceptions of formidability and prestige, especially in societies with higher homicide rates and greater relational mobility in which male intrasexual competition may be more intense and rapid identification of high-status competitors may be exigent. High female fo increased women's perceptions of flirtatiousness where relational mobility was lower and threats to mating relationships may be greater. These results indicate that the influence of fo on social perceptions depends on socioecological variables, including those related to competition for status and mates.


Asunto(s)
Voz , Adulto , Humanos , Masculino , Femenino , Homicidio , Percepción Social , Parejas Sexuales
8.
Psychol Sci ; 35(5): 543-557, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38620057

RESUMEN

Recently, gender-ambiguous (nonbinary) voices have been added to voice assistants to combat gender stereotypes and foster inclusion. However, if people react negatively to such voices, these laudable efforts may be counterproductive. In five preregistered studies (N = 3,684 adult participants) we found that people do react negatively, rating products described by narrators with gender-ambiguous voices less favorably than when they are described by clearly male or female narrators. The voices create a feeling of unease, or social disfluency, that affects evaluations of the products being described. These effects are best explained by low familiarity with voices that sound ambiguous. Thus, initial negative reactions can be overcome with more exposure.


Asunto(s)
Voz , Humanos , Femenino , Masculino , Adulto , Adulto Joven , Estereotipo , Percepción Social , Identidad de Género , Adolescente , Persona de Mediana Edad
9.
Ear Hear ; 45(4): 952-968, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38616318

RESUMEN

OBJECTIVES: Postlingually deaf adults with cochlear implants (CIs) have difficulties with perceiving differences in speakers' voice characteristics and benefit little from voice differences for the perception of speech in competing speech. However, not much is known yet about the perception and use of voice characteristics in prelingually deaf implanted children with CIs. Unlike CI adults, most CI children became deaf during the acquisition of language. Extensive neuroplastic changes during childhood could make CI children better at using the available acoustic cues than CI adults, or the lack of exposure to a normal acoustic speech signal could make it more difficult for them to learn which acoustic cues they should attend to. This study aimed to examine to what degree CI children can perceive voice cues and benefit from voice differences for perceiving speech in competing speech, comparing their abilities to those of normal-hearing (NH) children and CI adults. DESIGN: CI children's voice cue discrimination (experiment 1), voice gender categorization (experiment 2), and benefit from target-masker voice differences for perceiving speech in competing speech (experiment 3) were examined in three experiments. The main focus was on the perception of mean fundamental frequency (F0) and vocal-tract length (VTL), the primary acoustic cues related to speakers' anatomy and perceived voice characteristics, such as voice gender. RESULTS: CI children's F0 and VTL discrimination thresholds indicated lower sensitivity to differences compared with their NH-age-equivalent peers, but their mean discrimination thresholds of 5.92 semitones (st) for F0 and 4.10 st for VTL indicated higher sensitivity than postlingually deaf CI adults with mean thresholds of 9.19 st for F0 and 7.19 st for VTL. Furthermore, CI children's perceptual weighting of F0 and VTL cues for voice gender categorization closely resembled that of their NH-age-equivalent peers, in contrast with CI adults. Finally, CI children had more difficulties in perceiving speech in competing speech than their NH-age-equivalent peers, but they performed better than CI adults. Unlike CI adults, CI children showed a benefit from target-masker voice differences in F0 and VTL, similar to NH children. CONCLUSION: Although CI children's F0 and VTL voice discrimination scores were overall lower than those of NH children, their weighting of F0 and VTL cues for voice gender categorization and their benefit from target-masker differences in F0 and VTL resembled that of NH children. Together, these results suggest that prelingually deaf implanted CI children can effectively utilize spectrotemporally degraded F0 and VTL cues for voice and speech perception, generally outperforming postlingually deaf CI adults in comparable tasks. These findings underscore the presence of F0 and VTL cues in the CI signal to a certain degree and suggest other factors contributing to the perception challenges faced by CI adults.


Asunto(s)
Implantación Coclear , Implantes Cocleares , Señales (Psicología) , Sordera , Percepción del Habla , Humanos , Sordera/rehabilitación , Masculino , Femenino , Niño , Adulto , Adulto Joven , Adolescente , Voz/fisiología , Estudios de Casos y Controles , Preescolar , Persona de Mediana Edad
10.
Conscious Cogn ; 123: 103718, 2024 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-38880020

RESUMEN

The phenomenon of "hearing voices" can be found not only in psychotic disorders, but also in the general population, with individuals across cultures reporting auditory perceptions of supernatural beings. In our preregistered study, we investigated a possible mechanism of such experiences, grounded in the predictive processing model of agency detection. We predicted that in a signal detection task, expecting less or more voices than actually present would drive the response bias toward a more conservative and liberal response strategy, respectively. Moreover, we hypothesized that including sensory noise would enhance these expectancy effects. In line with our predictions, the findings show that detection of voices relies on expectations and that this effect is especially pronounced in the case of unreliable sensory data. As such, the study contributes to our understanding of the predictive processes in hearing and the building blocks of voice hearing experiences.


Asunto(s)
Percepción Auditiva , Detección de Señal Psicológica , Humanos , Femenino , Adulto , Masculino , Adulto Joven , Percepción Auditiva/fisiología , Detección de Señal Psicológica/fisiología , Voz/fisiología , Anticipación Psicológica/fisiología , Alucinaciones/fisiopatología
11.
Anesth Analg ; 139(3): 639-646, 2024 Sep 01.
Artículo en Inglés | MEDLINE | ID: mdl-39008420

RESUMEN

BACKGROUND: Anesthesia monitors and devices are usually controlled with some combination of dials, keypads, a keyboard, or a touch screen. Thus, anesthesiologists can operate their monitors only when they are physically close to them, and not otherwise task-loaded with sterile procedures such as line or block placement. Voice recognition technology has become commonplace and may offer advantages in anesthesia practice such as reducing surface contamination rates and allowing anesthesiologists to effect changes in monitoring and therapy when they would otherwise presently be unable to do so. We hypothesized that this technology is practicable and that anesthesiologists would consider it useful. METHODS: A novel voice-driven prototype controller was designed for the GE Solar 8000M anesthesia patient monitor. The apparatus was implemented using a Raspberry Pi 4 single-board computer, an external conference audio device, a Google Cloud Speech-to-Text platform, and a modified Solar controller to effect commands. Fifty anesthesia providers tested the prototype. Evaluations and surveys were completed in a nonclinical environment to avoid any ethical or safety concerns regarding the use of the device in direct patient care. All anesthesiologists sampled were fluent English speakers; many with inflections from their first language or national origin, reflecting diversity in the population of practicing anesthesiologists. RESULTS: The prototype was uniformly well-received by anesthesiologists. Ease-of-use, usefulness, and effectiveness were assessed on a Likert scale with means of 9.96, 7.22, and 8.48 of 10, respectively. No population cofactors were associated with these results. Advancing level of training (eg, nonattending versus attending) was not correlated with any preference. Accent of country or region was not correlated with any preference. Vocal pitch register did not correlate with any preference. Statistical analyses were performed with analysis of variance and the unpaired t -test. CONCLUSIONS: The use of voice recognition to control operating room monitors was well-received anesthesia providers. Additional commands are easily implemented on the prototype controller. No adverse relationship was found between acceptability and level of anesthesia experience, pitch of voice, or presence of accent. Voice recognition is a promising method of controlling anesthesia monitors and devices that could potentially increase usability and situational awareness in circumstances where the anesthesiologist is otherwise out-of-position or task-loaded.


Asunto(s)
Anestesiólogos , Monitoreo Intraoperatorio , Humanos , Monitoreo Intraoperatorio/instrumentación , Monitoreo Intraoperatorio/métodos , Masculino , Diseño de Equipo , Voz , Software de Reconocimiento del Habla , Femenino , Anestesiología/instrumentación , Persona de Mediana Edad , Presión Sanguínea , Anestesia , Determinación de la Presión Sanguínea/instrumentación , Determinación de la Presión Sanguínea/métodos , Adulto
12.
J Med Internet Res ; 26: e42850, 2024 Jan 11.
Artículo en Inglés | MEDLINE | ID: mdl-38206657

RESUMEN

BACKGROUND: Web-based health care has the potential to improve health care access and convenience for patients with limited mobility, but its success depends on active physician participation. The economic returns of internet-based health care initiatives are an important factor that can motivate physicians to continue their participation. Although several studies have examined the communication patterns and influences of web-based health consultations, the correlation between physicians' communication characteristics and their economic returns remains unexplored. OBJECTIVE: This study aims to investigate how the linguistic features of 2 modes of physician-patient communication, instrumental and affective, determine the physician's economic returns, measured by the honorarium their patients agree to pay per consultation. We also examined the moderating effects of communication media (web-based text messages and voice messages) and the compounding effects of different communication features on economic returns. METHODS: We collected 40,563 web-based consultations from 528 physicians across 4 disease specialties on a large, web-based health care platform in China. Communication features were extracted using linguistic inquiry and word count, and we used multivariable linear regression and K-means clustering to analyze the data. RESULTS: We found that the use of cognitive processing language (ie, words related to insight, causation, tentativeness, and certainty) in instrumental communication and positive emotion-related words in affective communication were positively associated with the economic returns of physicians. However, the extensive use of discrepancy-related words could generate adverse effects. We also found that the use of voice messages for service delivery magnified the effects of cognitive processing language but did not moderate the effects of affective processing language. The highest economic returns were associated with consultations in which the physicians used few expressions related to negative emotion; used more terms associated with positive emotions; and later, used instrumental communication language. CONCLUSIONS: Our study provides empirical evidence about the relationship between physicians' communication characteristics and their economic returns. It contributes to a better understanding of patient-physician interactions from a professional-client perspective and has practical implications for physicians and web-based health care platform executives.


Asunto(s)
Médicos , Voz , Humanos , Comunicación , Lingüística , Lenguaje
13.
J Med Internet Res ; 26: e47134, 2024 Jan 09.
Artículo en Inglés | MEDLINE | ID: mdl-38194260

RESUMEN

BACKGROUND: Embodied conversational agents (ECAs) are computer-generated animated humanlike characters that interact with users through verbal and nonverbal behavioral cues. They are increasingly used in a range of fields, including health care. OBJECTIVE: This scoping review aims to identify the current practice in the development and evaluation of ECAs for chronic diseases. METHODS: We applied a methodological framework in this review. A total of 6 databases (ie, PubMed, Embase, CINAHL, ACM Digital Library, IEEE Xplore Digital Library, and Web of Science) were searched using a combination of terms related to ECAs and health in October 2023. Two independent reviewers selected the studies and extracted the data. This review followed the PRISMA-ScR (Preferred Reporting Items of Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) statement. RESULTS: The literature search found 6332 papers, of which 36 (0.57%) met the inclusion criteria. Among the 36 studies, 27 (75%) originated from the United States, and 28 (78%) were published from 2020 onward. The reported ECAs covered a wide range of chronic diseases, with a focus on cancers, atrial fibrillation, and type 2 diabetes, primarily to promote screening and self-management. Most ECAs were depicted as middle-aged women based on screenshots and communicated with users through voice and nonverbal behavior. The most frequently reported evaluation outcomes were acceptability and effectiveness. CONCLUSIONS: This scoping review provides valuable insights for technology developers and health care professionals regarding the development and implementation of ECAs. It emphasizes the importance of technological advances in the embodiment, personalized strategy, and communication modality and requires in-depth knowledge of user preferences regarding appearance, animation, and intervention content. Future studies should incorporate measures of cost, efficiency, and productivity to provide a comprehensive evaluation of the benefits of using ECAs in health care.


Asunto(s)
Fibrilación Atrial , Diabetes Mellitus Tipo 2 , Voz , Persona de Mediana Edad , Humanos , Femenino , Comunicación , Enfermedad Crónica
14.
J Med Internet Res ; 26: e57258, 2024 Aug 07.
Artículo en Inglés | MEDLINE | ID: mdl-39110963

RESUMEN

BACKGROUND: The integration of smart technologies, including wearables and voice-activated devices, is increasingly recognized for enhancing the independence and well-being of older adults. However, the long-term dynamics of their use and the coadaptation process with older adults remain poorly understood. This scoping review explores how interactions between older adults and smart technologies evolve over time to improve both user experience and technology utility. OBJECTIVE: This review synthesizes existing research on the coadaptation between older adults and smart technologies, focusing on longitudinal changes in use patterns, the effectiveness of technological adaptations, and the implications for future technology development and deployment to improve user experiences. METHODS: Following the Joanna Briggs Institute Reviewer's Manual and PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) guidelines, this scoping review examined peer-reviewed papers from databases including Ovid MEDLINE, Ovid Embase, PEDro, Ovid PsycINFO, and EBSCO CINAHL from the year 2000 to August 28, 2023, and included forward and backward searches. The search was updated on March 1, 2024. Empirical studies were included if they involved (1) individuals aged 55 years or older living independently and (2) focused on interactions and adaptations between older adults and wearables and voice-activated virtual assistants in interventions for a minimum period of 8 weeks. Data extraction was informed by the selection and optimization with compensation framework and the sex- and gender-based analysis plus theoretical framework and used a directed content analysis approach. RESULTS: The search yielded 16,143 papers. Following title and abstract screening and a full-text review, 5 papers met the inclusion criteria. Study populations were mostly female participants and aged 73-83 years from the United States and engaged with voice-activated virtual assistants accessed through smart speakers and wearables. Users frequently used simple commands related to music and weather, integrating devices into daily routines. However, communication barriers often led to frustration due to devices' inability to recognize cues or provide personalized responses. The findings suggest that while older adults can integrate smart technologies into their lives, a lack of customization and user-friendly interfaces hinder long-term adoption and satisfaction. The studies highlight the need for technology to be further developed so they can better meet this demographic's evolving needs and call for research addressing small sample sizes and limited diversity. CONCLUSIONS: Our findings highlight a critical need for continued research into the dynamic and reciprocal relationship between smart technologies and older adults over time. Future studies should focus on more diverse populations and extend monitoring periods to provide deeper insights into the coadaptation process. Insights gained from this review are vital for informing the development of more intuitive, user-centric smart technology solutions to better support the aging population in maintaining independence and enhancing their quality of life. INTERNATIONAL REGISTERED REPORT IDENTIFIER (IRRID): RR2-10.2196/51129.


Asunto(s)
Dispositivos Electrónicos Vestibles , Humanos , Anciano , Persona de Mediana Edad , Femenino , Masculino , Anciano de 80 o más Años , Voz , Estudios Longitudinales
15.
Chaos ; 34(2)2024 Feb 01.
Artículo en Inglés | MEDLINE | ID: mdl-38386906

RESUMEN

In humans, ventricular folds are located superiorly to the vocal folds. Under special circumstances such as voice pathology or singing, they vibrate together with the vocal folds to contribute to the production of voice. In the present study, experimental data measured from physical models of the vocal and ventricular folds were analyzed in the light of nonlinear dynamics. The physical models provide a useful experimental framework to study the biomechanics of human vocalizations. Of particular interest in this experiment are co-oscillations of the vocal and ventricular folds, occasionally accompanied by irregular dynamics. We show that such a system can be regarded as two coupled oscillators, which give rise to various cooperative behaviors such as synchronized oscillations with a 1:1 or 1:2 frequency ratio and desynchronized oscillations with torus or chaos. The insight gained from the view of nonlinear dynamics should be of significant use for the diagnosis of voice pathologies, such as ventricular fold dysphonia.


Asunto(s)
Pliegues Vocales , Voz , Humanos , Dinámicas no Lineales , Fenómenos Biomecánicos
16.
Am J Otolaryngol ; 45(2): 104139, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38101138

RESUMEN

INTRODUCTION: Mutational voice disorder is the inability of the voice to adjust to the changes in the larynx during puberty, resulting in the speaking fundamental frequency failing to decrease. Standard treatments for mutational voice disorder are voice therapy and thyroplasty. However, voice therapy takes time to show its effects, and thyroplasty is highly invasive. Herein, we present a case of mutational voice disorder successfully treated with intracordal trafermin injection. CASE SUMMARY: A 31-year-old male patient was diagnosed with mutational voice disorder and offered standard treatment, but he requested a less invasive treatment with early effects. We performed intracordal trafermin injection with his consent. Two months after the procedure, the speaking fundamental frequency decreased from 155.5 Hz to 93.0 Hz, and the voice handicap index decreased from 14 to 2. DISCUSSION: This case suggests that intracordal trafermin injection is an effective treatment option for mutational voice disorder. Furthermore, compared with the standard treatment methods, it is less invasive and provides effects shortly with only one injection.


Asunto(s)
Factores de Crecimiento de Fibroblastos , Fragmentos de Péptidos , Trastornos de la Voz , Voz , Masculino , Humanos , Adulto , Trastornos de la Voz/tratamiento farmacológico , Trastornos de la Voz/cirugía , Resultado del Tratamiento , Inyecciones
17.
Am J Otolaryngol ; 45(2): 104174, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38101141

RESUMEN

OBJECTIVES: Vocal cord nodules (VCNs) are the most common cause of dysphonia in school-aged children, with potential negative impacts on quality of life including diminished self-esteem and academic performance. The standard of care for VCNs is conservative management which ranges from voice hygiene to speech therapy with a focus on voice otherwise known as voice therapy, with surgical excision reserved for refractory cases. Thus, few studies have analyzed outcomes of surgical management of VCNs. The purpose of this study is to assess the prevalence and efficacy of surgical excision of VCNs when compared to speech therapy. METHODS: Children with VCNs seen at a single tertiary care institution between 2015 and 2020 were identified by ICD-9 code 478.5 and ICD-10 code J38.2. Demographics, objective voice assessment, intervention, and follow-up assessment data were reviewed. Frequencies, medians, and interquartile ranges were calculated. Time to resolution and improvement were assessed by Cox proportional hazards model. Univariate logistic regression was performed. A P value of <0.05 was considered statistically significant. RESULTS: Three hundred sixty-eight patients diagnosed with VCNs were identified. 169 patients received intervention for VCNs, with 159 (43.2 %) receiving speech therapy alone and 5 (1.4 %) receiving surgery alone. On bivariate analysis, there was no significant difference in demographic features between treatment groups, however speech therapy patients did have a longer follow-up time. 154 patients underwent objective voice assessment at the time of VCN diagnosis. Among these patients, 95 (61.7 %) received speech therapy and 59 (40.3 %) received no intervention. Speech therapy patients had significantly higher pVHI scores, however there was no significant difference in CAPE-V Overall Severity scores or computerized voice assessment analysis. On Cox proportional hazards analysis, surgical intervention was associated with faster resolution and faster improvement of dysphonic symptoms. On binary logistic regression, surgery was associated with a significantly greater proportion of patients reporting resolution of dysphonic symptoms, however there was no significant difference in proportion of patients reporting improvement of dysphonia. CONCLUSION: For most patients with VCNs, conservative measures such as voice hygiene and speech therapy remain first line, however certain patients may benefit from the rapid improvement and resolution of symptoms that surgical intervention may provide.


Asunto(s)
Disfonía , Enfermedades de la Laringe , Pólipos , Voz , Niño , Humanos , Disfonía/etiología , Disfonía/diagnóstico , Pliegues Vocales/cirugía , Calidad de Vida , Enfermedades de la Laringe/diagnóstico
18.
Eur Arch Otorhinolaryngol ; 281(4): 1857-1864, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38183455

RESUMEN

OBJECTIVE: This study evaluated the swallowing and voice function of laryngeal cancer patients after Supracricoid Partial Laryngectomy(SCPL), and its influence on quality of life to provide a reference for the selection of surgical methods for laryngeal cancer patients. METHODS: Twenty-one patients who received SCPL between April 2015 and November 2021 were included. Each patient's swallowing function and quality of life were assessed through fiberoptic endoscopic examination of swallowing (FEES) and the M.D. Anderson Dysphagia Inventory (MDADI). Fundamental, jitter, shimmer, maximum phonation time (MPT), and voice handicap index-10 (VHI-10) were performed to assess voice function and voice-related quality of life. RESULTS: The results of the FEES of the 21 patients were as follows: the rates of pharyngeal residue after swallowing solid, semiliquid, and liquid food were 0%, 28.57%, and 38.09%, respectively; the rates of laryngeal infiltration after swallowing solid, semiliquid, and liquid food were 0%, 28.57%, and 4.76%, respectively; and aspiration did not occur in any of the patients. In the evaluation of swallowing quality of life, the mean total MDADI score was 92.6 ± 6.32. The voice function evaluation showed that the mean F0, jitter, shimmer, and MPT values were 156.01 ± 120.87 (HZ), 11.57 ± 6.21 (%), 35.37 ± 14.16 (%) and 7.85 ± 6.08 (s), respectively. The mean total VHI-10 score was 7.14 ± 4.84. CONCLUSION: SCPL provides patients with satisfactory swallowing and voice function. The patients in this study were satisfied with their quality of life in terms of swallowing and voice. SCPL can be used as a surgical method to preserve laryngeal function in patients with laryngeal cancer.


Asunto(s)
Neoplasias Laríngeas , Voz , Humanos , Laringectomía/efectos adversos , Laringectomía/métodos , Deglución , Neoplasias Laríngeas/cirugía , Calidad de Vida
19.
Eur Arch Otorhinolaryngol ; 281(5): 2707-2716, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38319369

RESUMEN

PURPOSE: This cross-sectional study aimed to investigate the potential of voice analysis as a prescreening tool for type II diabetes mellitus (T2DM) by examining the differences in voice recordings between non-diabetic and T2DM participants. METHODS: 60 participants diagnosed as non-diabetic (n = 30) or T2DM (n = 30) were recruited on the basis of specific inclusion and exclusion criteria in Iran between February 2020 and September 2023. Participants were matched according to their year of birth and then placed into six age categories. Using the WhatsApp application, participants recorded the translated versions of speech elicitation tasks. Seven acoustic features [fundamental frequency, jitter, shimmer, harmonic-to-noise ratio (HNR), cepstral peak prominence (CPP), voice onset time (VOT), and formant (F1-F2)] were extracted from each recording and analyzed using Praat software. Data was analyzed with Kolmogorov-Smirnov, two-way ANOVA, post hoc Tukey, binary logistic regression, and student t tests. RESULTS: The comparison between groups showed significant differences in fundamental frequency, jitter, shimmer, CPP, and HNR (p < 0.05), while there were no significant differences in formant and VOT (p > 0.05). Binary logistic regression showed that shimmer was the most significant predictor of the disease group. There was also a significant difference between diabetes status and age, in the case of CPP. CONCLUSIONS: Participants with type II diabetes exhibited significant vocal variations compared to non-diabetic controls.


Asunto(s)
Diabetes Mellitus Tipo 2 , Voz , Humanos , Calidad de la Voz , Acústica del Lenguaje , Diabetes Mellitus Tipo 2/complicaciones , Estudios Transversales , Medición de la Producción del Habla , Acústica
20.
Eur Arch Otorhinolaryngol ; 281(5): 2499-2505, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38365991

RESUMEN

PURPOSE: Arytenoid adduction as an addition to medialisation thyroplasty is highly advocated by some surgeons in selected cases but deemed less necessary by others in patients with unilateral vocal fold paralysis. This study aims to evaluate the additional benefits on voice outcome of arytenoid adduction in patients with unilateral vocal fold paralysis undergoing medialisation thyroplasty using intra-operative voice measurements. DESIGN/METHODS: A prospective study was conducted. Voice audio recordings were obtained at 4 moments; 1. direct prior to the start of surgery, 2. during surgery after medialisation thyroplasty, 3. during surgery after medialisation and arytenoid adduction, 3 months postoperative. At these same timepoints patients rated their own voice on a numeric rating scale between 0 and 10. The blinded recordings were rated by consensus in a team of experienced listeners, using the Grade of the GRBAS scale. Furthermore, the Voice Handicap Index was administered before and at 3 months after surgery. RESULTS: Ten patients who underwent medialisation and arytenoid adduction at our tertiary referral hospital between 2021 and 2022, were included. One patient was excluded after surgery. The intraoperative measurements showed a Grade score of 1.4 preoperatively, improving to 1.2 after medialisation, 1.2 after medialisation and arytenoid adduction, and further improving to 0.4 at 3 months postoperative, which was a not statistically significant improvement (p = 0.2). The intraoperative subjective numeric rating scale showed a statistically significant improvement from 3.9 preoperatively, to 6.1 after medialisation, 7.1 after medialisation and arytenoid adduction and a 7.6 at 3 months postoperative (p = 0.001). The Voice Handicap Index total score showed a statistically significant improvement from 71 points before surgery to 13 at 3 months after surgery (p = 0.008). CONCLUSIONS: Our study using intraoperative voice measurements indicate that the addition of arytenoid adduction to medialisation thyroplasty is a benefit in selected patients although more studies are needed due to the many limitations inherent to this field of investigation.


Asunto(s)
Laringoplastia , Parálisis de los Pliegues Vocales , Voz , Humanos , Estudios Prospectivos , Calidad de la Voz , Parálisis de los Pliegues Vocales/cirugía , Cartílago Aritenoides/cirugía , Resultado del Tratamiento
SELECCIÓN DE REFERENCIAS
Detalles de la búsqueda