Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 8.377
Filtrar
1.
Sci Rep ; 14(1): 15611, 2024 Jul 06.
Artigo em Inglês | MEDLINE | ID: mdl-38971806

RESUMO

This study compares how English-speaking adults and children from the United States adapt their speech when talking to a real person and a smart speaker (Amazon Alexa) in a psycholinguistic experiment. Overall, participants produced more effortful speech when talking to a device (longer duration and higher pitch). These differences also varied by age: children produced even higher pitch in device-directed speech, suggesting a stronger expectation to be misunderstood by the system. In support of this, we see that after a staged recognition error by the device, children increased pitch even more. Furthermore, both adults and children displayed the same degree of variation in their responses for whether "Alexa seems like a real person or not", further indicating that children's conceptualization of the system's competence shaped their register adjustments, rather than an increased anthropomorphism response. This work speaks to models on the mechanisms underlying speech production, and human-computer interaction frameworks, providing support for routinized theories of spoken interaction with technology.


Assuntos
Fala , Humanos , Adulto , Criança , Masculino , Feminino , Fala/fisiologia , Adulto Jovem , Adolescente , Psicolinguística
2.
Sensors (Basel) ; 24(13)2024 Jun 25.
Artigo em Inglês | MEDLINE | ID: mdl-39000889

RESUMO

Emotions in speech are expressed in various ways, and the speech emotion recognition (SER) model may perform poorly on unseen corpora that contain different emotional factors from those expressed in training databases. To construct an SER model robust to unseen corpora, regularization approaches or metric losses have been studied. In this paper, we propose an SER method that incorporates relative difficulty and labeling reliability of each training sample. Inspired by the Proxy-Anchor loss, we propose a novel loss function which gives higher gradients to the samples for which the emotion labels are more difficult to estimate among those in the given minibatch. Since the annotators may label the emotion based on the emotional expression which resides in the conversational context or other modality but is not apparent in the given speech utterance, some of the emotional labels may not be reliable and these unreliable labels may affect the proposed loss function more severely. In this regard, we propose to apply label smoothing for the samples misclassified by a pre-trained SER model. Experimental results showed that the performance of the SER on unseen corpora was improved by adopting the proposed loss function with label smoothing on the misclassified data.


Assuntos
Emoções , Fala , Humanos , Emoções/fisiologia , Fala/fisiologia , Algoritmos , Reprodutibilidade dos Testes , Reconhecimento Automatizado de Padrão/métodos , Bases de Dados Factuais
3.
J Robot Surg ; 18(1): 287, 2024 Jul 18.
Artigo em Inglês | MEDLINE | ID: mdl-39026112

RESUMO

Transoral robotic surgery (TORS) has been introduced to head and neck surgery as a minimally invasive techqniques to improve the functional outcomes of patients. Compare the functional outcome for swallowing and speech in each site of TORS within the head and neck. Retrospective cohort study for patients who underwent TORS within the head and neck unit. Patients were assessed at four different time points (one day, one month, six months and twelve months, respectively) with bedside/office testing. Methods of testing for swallowing assessment were by the International Dysphagia Diet Standardization Initiative (IDDSI), and speech assessments were carried out using the Understandability of Speech score (USS). Outcomes were compared to patient-specific pre-treatment baseline levels. 68 patients were included. 75% and 40% of the patients resumed normal fluid intake and normal diet immediately after surgery. 8.8% required a temporary feeding tube, with 1% required gastrostomy. There was a steep improvement in diet between 3 and 6 months. Fluid and diet consistency dropped significantly following the majority of transoral robotic surgery with more noticeable diet changes. Early deterioration in diet is temporary and manageable with a modified diet. Rapid recovery of swallowing is achieved before the first year. There is no long-term effect on speech.


Assuntos
Transtornos de Deglutição , Deglutição , Procedimentos Cirúrgicos Robóticos , Fala , Humanos , Procedimentos Cirúrgicos Robóticos/métodos , Deglutição/fisiologia , Masculino , Feminino , Estudos Retrospectivos , Fala/fisiologia , Pessoa de Meia-Idade , Idoso , Transtornos de Deglutição/etiologia , Resultado do Tratamento , Boca , Adulto , Neoplasias de Cabeça e Pescoço/cirurgia , Idoso de 80 Anos ou mais
4.
Elife ; 132024 Jul 22.
Artigo em Inglês | MEDLINE | ID: mdl-39038076

RESUMO

To what extent does speech and music processing rely on domain-specific and domain-general neural networks? Using whole-brain intracranial EEG recordings in 18 epilepsy patients listening to natural, continuous speech or music, we investigated the presence of frequency-specific and network-level brain activity. We combined it with a statistical approach in which a clear operational distinction is made between shared, preferred, and domain-selective neural responses. We show that the majority of focal and network-level neural activity is shared between speech and music processing. Our data also reveal an absence of anatomical regional selectivity. Instead, domain-selective neural responses are restricted to distributed and frequency-specific coherent oscillations, typical of spectral fingerprints. Our work highlights the importance of considering natural stimuli and brain dynamics in their full complexity to map cognitive and brain functions.


Assuntos
Música , Humanos , Masculino , Feminino , Adulto , Rede Nervosa/fisiologia , Fala/fisiologia , Percepção Auditiva/fisiologia , Epilepsia/fisiopatologia , Adulto Jovem , Eletroencefalografia , Córtex Cerebral/fisiologia , Eletrocorticografia , Percepção da Fala/fisiologia , Pessoa de Meia-Idade , Mapeamento Encefálico
5.
Hum Brain Mapp ; 45(11): e26797, 2024 Aug 01.
Artigo em Inglês | MEDLINE | ID: mdl-39041175

RESUMO

Speech comprehension is crucial for human social interaction, relying on the integration of auditory and visual cues across various levels of representation. While research has extensively studied multisensory integration (MSI) using idealised, well-controlled stimuli, there is a need to understand this process in response to complex, naturalistic stimuli encountered in everyday life. This study investigated behavioural and neural MSI in neurotypical adults experiencing audio-visual speech within a naturalistic, social context. Our novel paradigm incorporated a broader social situational context, complete words, and speech-supporting iconic gestures, allowing for context-based pragmatics and semantic priors. We investigated MSI in the presence of unimodal (auditory or visual) or complementary, bimodal speech signals. During audio-visual speech trials, compared to unimodal trials, participants more accurately recognised spoken words and showed a more pronounced suppression of alpha power-an indicator of heightened integration load. Importantly, on the neural level, these effects surpassed mere summation of unimodal responses, suggesting non-linear MSI mechanisms. Overall, our findings demonstrate that typically developing adults integrate audio-visual speech and gesture information to facilitate speech comprehension in noisy environments, highlighting the importance of studying MSI in ecologically valid contexts.


Assuntos
Gestos , Percepção da Fala , Humanos , Feminino , Masculino , Percepção da Fala/fisiologia , Adulto Jovem , Adulto , Percepção Visual/fisiologia , Eletroencefalografia , Compreensão/fisiologia , Estimulação Acústica , Fala/fisiologia , Encéfalo/fisiologia , Estimulação Luminosa/métodos
6.
Sci Rep ; 14(1): 15787, 2024 Jul 09.
Artigo em Inglês | MEDLINE | ID: mdl-38982177

RESUMO

Diagnostic tests for Parkinsonism based on speech samples have shown promising results. Although abnormal auditory feedback integration during speech production and impaired rhythmic organization of speech are known in Parkinsonism, these aspects have not been incorporated into diagnostic tests. This study aimed to identify Parkinsonism using a novel speech behavioral test that involved rhythmically repeating syllables under different auditory feedback conditions. The study included 30 individuals with Parkinson's disease (PD) and 30 healthy subjects. Participants were asked to rhythmically repeat the PA-TA-KA syllable sequence, both whispering and speaking aloud under various listening conditions. The results showed that individuals with PD had difficulties in whispering and articulating under altered auditory feedback conditions, exhibited delayed speech onset, and demonstrated inconsistent rhythmic structure across trials compared to controls. These parameters were then fed into a supervised machine-learning algorithm to differentiate between the two groups. The algorithm achieved an accuracy of 85.4%, a sensitivity of 86.5%, and a specificity of 84.3%. This pilot study highlights the potential of the proposed behavioral paradigm as an objective and accessible (both in cost and time) test for identifying individuals with Parkinson's disease.


Assuntos
Retroalimentação Sensorial , Doença de Parkinson , Fala , Humanos , Feminino , Masculino , Idoso , Doença de Parkinson/fisiopatologia , Doença de Parkinson/diagnóstico , Pessoa de Meia-Idade , Fala/fisiologia , Retroalimentação Sensorial/fisiologia , Projetos Piloto , Transtornos Parkinsonianos/fisiopatologia , Estudos de Casos e Controles
7.
Commun Biol ; 7(1): 818, 2024 Jul 05.
Artigo em Inglês | MEDLINE | ID: mdl-38969758

RESUMO

Speech brain-computer interfaces aim to support communication-impaired patients by translating neural signals into speech. While impressive progress was achieved in decoding performed, perceived and attempted speech, imagined speech remains elusive, mainly due to the absence of behavioral output. Nevertheless, imagined speech is advantageous since it does not depend on any articulator movements that might become impaired or even lost throughout the stages of a neurodegenerative disease. In this study, we analyzed electrocortigraphy data recorded from 16 participants in response to 3 speech modes: performed, perceived (listening), and imagined speech. We used a linear model to detect speech events and examined the contributions of each frequency band, from delta to high gamma, given the speech mode and electrode location. For imagined speech detection, we observed a strong contribution of gamma bands in the motor cortex, whereas lower frequencies were more prominent in the temporal lobe, in particular of the left hemisphere. Based on the similarities in frequency patterns, we were able to transfer models between speech modes and participants with similar electrode locations.


Assuntos
Interfaces Cérebro-Computador , Eletrocorticografia , Imaginação , Fala , Humanos , Eletrocorticografia/métodos , Fala/fisiologia , Masculino , Feminino , Adulto , Imaginação/fisiologia , Adulto Jovem , Córtex Motor/fisiologia
8.
Sci Rep ; 14(1): 16409, 2024 Jul 16.
Artigo em Inglês | MEDLINE | ID: mdl-39013983

RESUMO

A fundamental aspect of language processing is inferring others' minds from subtle variations in speech. The same word or sentence can often convey different meanings depending on its tempo, timing, and intonation-features often referred to as prosody. Although autistic children and adults are known to experience difficulty in making such inferences, the science remains unclear as to why. We hypothesize that detail-oriented perception in autism may interfere with the inference process if it lacks the adaptivity required to cope with the variability ubiquitous in human speech. Using a novel prosodic continuum that shifts the sentence meaning gradiently from a statement (e.g., "It's raining") to a question (e.g., "It's raining?"), we have investigated the perception and adaptation of receptive prosody in autistic adolescents and two groups of non-autistic controls. Autistic adolescents showed attenuated adaptivity in categorizing prosody, whereas they were equivalent to controls in terms of discrimination accuracy. Combined with recent findings in segmental (e.g., phoneme) recognition, the current results provide the basis for an emerging research framework for attenuated flexibility and reduced influence of contextual feedback as a possible source of deficits that hinder linguistic and social communication in autism.


Assuntos
Transtorno Autístico , Percepção da Fala , Humanos , Adolescente , Masculino , Feminino , Percepção da Fala/fisiologia , Transtorno Autístico/fisiopatologia , Transtorno Autístico/psicologia , Idioma , Criança , Fala/fisiologia
9.
PLoS One ; 19(7): e0305657, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39018339

RESUMO

Technological developments over the past few decades have changed the way people communicate, with platforms like social media and blogs becoming vital channels for international conversation. Even though hate speech is vigorously suppressed on social media, it is still a concern that needs to be constantly recognized and observed. The Arabic language poses particular difficulties in the detection of hate speech, despite the considerable efforts made in this area for English-language social media content. Arabic calls for particular consideration when it comes to hate speech detection because of its many dialects and linguistic nuances. Another degree of complication is added by the widespread practice of "code-mixing," in which users merge various languages smoothly. Recognizing this research vacuum, the study aims to close it by examining how well machine learning models containing variation features can detect hate speech, especially when it comes to Arabic tweets featuring code-mixing. Therefore, the objective of this study is to assess and compare the effectiveness of different features and machine learning models for hate speech detection on Arabic hate speech and code-mixing hate speech datasets. To achieve the objectives, the methodology used includes data collection, data pre-processing, feature extraction, the construction of classification models, and the evaluation of the constructed classification models. The findings from the analysis revealed that the TF-IDF feature, when employed with the SGD model, attained the highest accuracy, reaching 98.21%. Subsequently, these results were contrasted with outcomes from three existing studies, and the proposed method outperformed them, underscoring the significance of the proposed method. Consequently, our study carries practical implications and serves as a foundational exploration in the realm of automated hate speech detection in text.


Assuntos
Idioma , Aprendizado de Máquina , Mídias Sociais , Humanos , Fala/fisiologia
10.
Sci Rep ; 14(1): 16603, 2024 Jul 18.
Artigo em Inglês | MEDLINE | ID: mdl-39025957

RESUMO

Electrophysiological brain activity has been shown to synchronize with the quasi-regular repetition of grammatical phrases in connected speech-so-called phrase-rate neural tracking. Current debate centers around whether this phenomenon is best explained in terms of the syntactic properties of phrases or in terms of syntax-external information, such as the sequential repetition of parts of speech. As these two factors were confounded in previous studies, much of the literature is compatible with both accounts. Here, we used electroencephalography (EEG) to determine if and when the brain is sensitive to both types of information. Twenty native speakers of Mandarin Chinese listened to isochronously presented streams of monosyllabic words, which contained either grammatical two-word phrases (e.g., catch fish, sell house) or non-grammatical word combinations (e.g., full lend, bread far). Within the grammatical conditions, we varied two structural factors: the position of the head of each phrase and the type of attachment. Within the non-grammatical conditions, we varied the consistency with which parts of speech were repeated. Tracking was quantified through evoked power and inter-trial phase coherence, both derived from the frequency-domain representation of EEG responses. As expected, neural tracking at the phrase rate was stronger in grammatical sequences than in non-grammatical sequences without syntactic structure. Moreover, it was modulated by both attachment type and head position, revealing the structure-sensitivity of phrase-rate tracking. We additionally found that the brain tracks the repetition of parts of speech in non-grammatical sequences. These data provide an integrative perspective on the current debate about neural tracking effects, revealing that the brain utilizes regularities computed over multiple levels of linguistic representation in guiding rhythmic computation.


Assuntos
Encéfalo , Eletroencefalografia , Humanos , Masculino , Feminino , Adulto , Encéfalo/fisiologia , Adulto Jovem , Idioma , Percepção da Fala/fisiologia , Fala/fisiologia
11.
PLoS One ; 19(7): e0306272, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39028710

RESUMO

Abnormal speech prosody has been widely reported in individuals with autism. Many studies on children and adults with autism spectrum disorder speaking a non-tonal language showed deficits in using prosodic cues to mark focus. However, focus marking by autistic children speaking a tonal language is rarely examined. Cantonese-speaking children may face additional difficulties because tonal languages require them to use prosodic cues to achieve multiple functions simultaneously such as lexical contrasting and focus marking. This study bridges this research gap by acoustically evaluating the use of Cantonese speech prosody to mark information structure by Cantonese-speaking children with and without autism spectrum disorder. We designed speech production tasks to elicit natural broad and narrow focus production among these children in sentences with different tone combinations. Acoustic correlates of prosodic focus marking like f0, duration and intensity of each syllable were analyzed to examine the effect of participant group, focus condition and lexical tones. Our results showed differences in focus marking patterns between Cantonese-speaking children with and without autism spectrum disorder. The autistic children not only showed insufficient on-focus expansion in terms of f0 range and duration when marking focus, but also produced less distinctive tone shapes in general. There was no evidence that the prosodic complexity (i.e. sentences with single tones or combinations of tones) significantly affected focus marking in these autistic children and their typically-developing (TD) peers.


Assuntos
Transtorno do Espectro Autista , Idioma , Humanos , Transtorno do Espectro Autista/fisiopatologia , Transtorno do Espectro Autista/psicologia , Masculino , Feminino , Criança , Acústica da Fala , Pré-Escolar , Fala/fisiologia
12.
Dental Press J Orthod ; 29(3): e2423277, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38985077

RESUMO

OBJECTIVE: This study aimed to compare the influence of four different maxillary removable orthodontic retainers on speech. MATERIAL AND METHODS: Eligibility criteria for sample selection were: 20-40-year subjects with acceptable occlusion, native speakers of Portuguese. The volunteers (n=21) were divided in four groups randomized with a 1:1:1:1 allocation ratio. The four groups used, in random order, the four types of retainers full-time for 21 days each, with a washout period of 7-days. The removable maxillary retainers were: conventional wraparound, wraparound with an anterior hole, U-shaped wraparound, and thermoplastic retainer. Three volunteers were excluded. The final sample comprised 18 subjects (11 male; 7 female) with mean age of 27.08 years (SD=4.65). The speech evaluation was performed in vocal excerpts recordings made before, immediately after, and 21 days after the installation of each retainer, with auditory-perceptual and acoustic analysis of formant frequencies F1 and F2 of the vowels. Repeated measures ANOVA and Friedman with Tukey tests were used for statistical comparison. RESULTS: Speech changes increased immediately after conventional wraparound and thermoplastic retainer installation, and reduced after 21 days, but not to normal levels. However, this increase was statistically significant only for the wraparound with anterior hole and the thermoplastic retainer. Formant frequencies of vowels were altered at initial time, and the changes remained in conventional, U-shaped and thermoplastic appliances after three weeks. CONCLUSIONS: The thermoplastic retainer was more harmful to the speech than wraparound appliances. The conventional and U-shaped retainers interfered less in speech. The three-week period was not sufficient for speech adaptation.


Assuntos
Estudos Cross-Over , Contenções Ortodônticas , Humanos , Feminino , Masculino , Adulto , Desenho de Aparelho Ortodôntico , Adulto Jovem , Fala/fisiologia
13.
PLoS One ; 19(6): e0305333, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38889141

RESUMO

Language is rooted in our ability to compose: We link words together, fusing their meanings. Links are not limited to neighboring words but often span intervening words. The ability to process these non-adjacent dependencies (NADs) conflicts with the brain's sampling of speech: We consume speech in chunks that are limited in time, containing only a limited number of words. It is unknown how we link words together that belong to separate chunks. Here, we report that we cannot-at least not so well. In our electroencephalography (EEG) study, 37 human listeners learned chunks and dependencies from an artificial grammar (AG) composed of syllables. Multi-syllable chunks to be learned were equal-sized, allowing us to employ a frequency-tagging approach. On top of chunks, syllable streams contained NADs that were either confined to a single chunk or crossed a chunk boundary. Frequency analyses of the EEG revealed a spectral peak at the chunk rate, showing that participants learned the chunks. NADs that cross boundaries were associated with smaller electrophysiological responses than within-chunk NADs. This shows that NADs are processed readily when they are confined to the same chunk, but not as well when crossing a chunk boundary. Our findings help to reconcile the classical notion that language is processed incrementally with recent evidence for discrete perceptual sampling of speech. This has implications for language acquisition and processing as well as for the general view of syntax in human language.


Assuntos
Eletroencefalografia , Idioma , Humanos , Feminino , Masculino , Adulto , Adulto Jovem , Percepção da Fala/fisiologia , Fala/fisiologia , Aprendizagem/fisiologia , Encéfalo/fisiologia
14.
Curr Opin Otolaryngol Head Neck Surg ; 32(4): 282-285, 2024 Aug 01.
Artigo em Inglês | MEDLINE | ID: mdl-38869616

RESUMO

PURPOSE OF REVIEW: The purpose of this review is to examine the current research of the posterior tongue tie and how it relates to breast feeding, solid feeding, and speech. RECENT FINDINGS: Recent findings show that the posterior tongue tie may play a role in effective breast feeding. SUMMARY: Ankyloglossia is the term used for the restriction of the movement of the tongue that impairs certain functions such as breastfeeding or bottle feeding, feeding with solids, and speech. Cadaver studies have shown that there can be a restriction of the tongue and oral tissues in some people relative to others. In some breast-feeding studies, releasing the posterior tie has been shown to improve certain aspects of tongue movement. There is little evidence for or against posterior tongue ties contributing to other problems such as speech and solid feeding. This article goes into depth about the current studies on posterior ankyloglossia.


Assuntos
Anquiloglossia , Aleitamento Materno , Língua , Humanos , Fala/fisiologia
15.
J Neural Eng ; 21(3)2024 Jun 27.
Artigo em Inglês | MEDLINE | ID: mdl-38885688

RESUMO

Objective.Brain-computer interfaces (BCIs) are technologies that bypass damaged or disrupted neural pathways and directly decode brain signals to perform intended actions. BCIs for speech have the potential to restore communication by decoding the intended speech directly. Many studies have demonstrated promising results using invasive micro-electrode arrays and electrocorticography. However, the use of stereo-electroencephalography (sEEG) for speech decoding has not been fully recognized.Approach.In this research, recently released sEEG data were used to decode Dutch words spoken by epileptic participants. We decoded speech waveforms from sEEG data using advanced deep-learning methods. Three methods were implemented: a linear regression method, an recurrent neural network (RNN)-based sequence-to-sequence model (RNN), and a transformer model.Main results.Our RNN and transformer models outperformed the linear regression significantly, while no significant difference was found between the two deep-learning methods. Further investigation on individual electrodes showed that the same decoding result can be obtained using only a few of the electrodes.Significance.This study demonstrated that decoding speech from sEEG signals is possible, and the location of the electrodes is critical to the decoding performance.


Assuntos
Interfaces Cérebro-Computador , Aprendizado Profundo , Eletroencefalografia , Fala , Humanos , Eletroencefalografia/métodos , Fala/fisiologia , Masculino , Feminino , Epilepsia/fisiopatologia , Epilepsia/diagnóstico , Técnicas Estereotáxicas , Adulto , Redes Neurais de Computação
16.
J Alzheimers Dis ; 100(1): 1-27, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38848181

RESUMO

Background: Dementia is a general term for several progressive neurodegenerative disorders including Alzheimer's disease. Timely and accurate detection is crucial for early intervention. Advancements in artificial intelligence present significant potential for using machine learning to aid in early detection. Objective: Summarize the state-of-the-art machine learning-based approaches for dementia prediction, focusing on non-invasive methods, as the burden on the patients is lower. Specifically, the analysis of gait and speech performance can offer insights into cognitive health through clinically cost-effective screening methods. Methods: A systematic literature review was conducted following the PRISMA protocol (Preferred Reporting Items for Systematic Reviews and Meta-Analyses). The search was performed on three electronic databases (Scopus, Web of Science, and PubMed) to identify the relevant studies published between 2017 to 2022. A total of 40 papers were selected for review. Results: The most common machine learning methods employed were support vector machine followed by deep learning. Studies suggested the use of multimodal approaches as they can provide comprehensive and better prediction performance. Deep learning application in gait studies is still in the early stages as few studies have applied it. Moreover, including features of whole body movement contribute to better classification accuracy. Regarding speech studies, the combination of different parameters (acoustic, linguistic, cognitive testing) produced better results. Conclusions: The review highlights the potential of machine learning, particularly non-invasive approaches, in the early prediction of dementia. The comparable prediction accuracies of manual and automatic speech analysis indicate an imminent fully automated approach for dementia detection.


Assuntos
Demência , Aprendizado de Máquina , Fala , Humanos , Demência/diagnóstico , Fala/fisiologia , Análise da Marcha/métodos
17.
Sci Rep ; 14(1): 14698, 2024 06 26.
Artigo em Inglês | MEDLINE | ID: mdl-38926416

RESUMO

Accommodating talker variability is a complex and multi-layered cognitive process. It involves shifting attention to the vocal characteristics of the talker as well as the linguistic content of their speech. Due to an interdependence between voice and phonological processing, multi-talker environments typically incur additional processing costs compared to single-talker environments. A failure or inability to efficiently distribute attention over multiple acoustic cues in the speech signal may have detrimental language learning consequences. Yet, no studies have examined effects of multi-talker processing in populations with atypical perceptual, social and language processing for communication, including autistic people. Employing a classic word-monitoring task, we investigated effects of talker variability in Australian English autistic (n = 24) and non-autistic (n = 28) adults. Listeners responded to target words (e.g., apple, duck, corn) in randomised sequences of words. Half of the sequences were spoken by a single talker and the other half by multiple talkers. Results revealed that autistic participants' sensitivity scores to accurately-spotted target words did not differ to those of non-autistic participants, regardless of whether they were spoken by a single or multiple talkers. As expected, the non-autistic group showed the well-established processing cost associated with talker variability (e.g., slower response times). Remarkably, autistic listeners' response times did not differ across single- or multi-talker conditions, indicating they did not show perceptual processing costs when accommodating talker variability. The present findings have implications for theories of autistic perception and speech and language processing.


Assuntos
Transtorno Autístico , Percepção da Fala , Humanos , Masculino , Feminino , Adulto , Percepção da Fala/fisiologia , Transtorno Autístico/fisiopatologia , Transtorno Autístico/psicologia , Adulto Jovem , Tempo de Reação/fisiologia , Fala/fisiologia , Atenção/fisiologia , Pessoa de Meia-Idade , Idioma
18.
Stomatologiia (Mosk) ; 103(3): 11-15, 2024.
Artigo em Russo | MEDLINE | ID: mdl-38904553

RESUMO

OBJECTIVE: To develop a universal system for assessing the speech function in patients with congenital palatal cleft in the postoperative period. MATERIALS AND METHODS: A universal system for assessing the speech function for patients with a palatal cleft can be applied both after the primary operation of uranoplasty and for patients diagnosed with velopharyngeal insufficiency (VPI). The patient's speech is assessed according to the following criteria: defects in the pronunciation of consonants by place of articulation: labial, labiodental, lingual-dental, lingual-palatal, lingual-alveolar; speech breathing; tongue position; directed air stream; voicing disorders; The patient's is also evaluated for the following findings: hypernasality (reflected speech); hypernasality (spontaneous speech); hyponasality; pharyngeal reflex; audible nasal emission/turbulence; facial grimaces; speech intelligibility. The speech therapy and dental assessments are added to obtain a value characterizing the patient's condition: from 0 to 10 scoring indicates than only speech therapy correction is needed; from 11 to 18 - the decision on the necessity of surgical treatment is made by the surgeon together with the speech therapist, from 18 to 25 - surgical treatment is necessary with subsequent sessions with a speech therapist. RESULTS: With the help of this questionnaire, the operating surgeon can more accurately and objectively assess in dynamics the result of the surgical treatment, regardless of the results of speech therapy treatment in the postoperative period. The creation of this scoring system for speech assessment is aimed at objectivizing the results of uranoplasty and speech-improving operations. It allows the surgeon to compare the effectiveness of different surgical methods. CONCLUSION: The universal scoring system for assessing the state of speech function can be applied in the diagnosis of a patient with a palatal cleft both after the primary operation on the palate and after corrective surgical interventions. It allows monitoring progress and identifying dynamics in surgical and speech therapy treatment.


Assuntos
Fissura Palatina , Insuficiência Velofaríngea , Humanos , Fissura Palatina/cirurgia , Fissura Palatina/complicações , Insuficiência Velofaríngea/cirurgia , Insuficiência Velofaríngea/fisiopatologia , Masculino , Feminino , Distúrbios da Fala/etiologia , Distúrbios da Fala/reabilitação , Distúrbios da Fala/diagnóstico , Fonoterapia/métodos , Criança , Adolescente , Fala/fisiologia , Inquéritos e Questionários , Período Pós-Operatório , Inteligibilidade da Fala
19.
Commun Biol ; 7(1): 748, 2024 Jun 20.
Artigo em Inglês | MEDLINE | ID: mdl-38902370

RESUMO

Human language relies on the correct processing of syntactic information, as it is essential for successful communication between speakers. As an abstract level of language, syntax has often been studied separately from the physical form of the speech signal, thus often masking the interactions that can promote better syntactic processing in the human brain. However, behavioral and neural evidence from adults suggests the idea that prosody and syntax interact, and studies in infants support the notion that prosody assists language learning. Here we analyze a MEG dataset to investigate how acoustic cues, specifically prosody, interact with syntactic representations in the brains of native English speakers. More specifically, to examine whether prosody enhances the cortical encoding of syntactic representations, we decode syntactic phrase boundaries directly from brain activity, and evaluate possible modulations of this decoding by the prosodic boundaries. Our findings demonstrate that the presence of prosodic boundaries improves the neural representation of phrase boundaries, indicating the facilitative role of prosodic cues in processing abstract linguistic features. This work has implications for interactive models of how the brain processes different linguistic features. Future research is needed to establish the neural underpinnings of prosody-syntax interactions in languages with different typological characteristics.


Assuntos
Idioma , Percepção da Fala , Humanos , Percepção da Fala/fisiologia , Masculino , Feminino , Adulto , Encéfalo/fisiologia , Fala/fisiologia , Adulto Jovem , Magnetoencefalografia , Linguística , Sinais (Psicologia)
20.
Curr Biol ; 34(12): 2719-2727.e5, 2024 Jun 17.
Artigo em Inglês | MEDLINE | ID: mdl-38823382

RESUMO

Turn-taking is a central feature of conversation across languages and cultures.1,2,3,4 This key social behavior requires numerous sensorimotor and cognitive operations1,5,6 that can be organized into three general phases: comprehension of a partner's turn, preparation of a speaker's own turn, and execution of that turn. Using intracranial electrocorticography, we recently demonstrated that neural activity related to these phases is functionally distinct during turn-taking.7 In particular, networks active during the perceptual and articulatory stages of turn-taking consisted of structures known to be important for speech-related sensory and motor processing,8,9,10,11,12,13,14,15,16,17 while putative planning dynamics were most regularly observed in the caudal inferior frontal gyrus (cIFG) and the middle frontal gyrus (cMFG). To test if these structures are necessary for planning during spoken interaction, we used direct electrical stimulation (DES) to transiently perturb cortical function in neurosurgical patient-volunteers performing a question-answer task.7,18,19 We found that stimulating the cIFG and cMFG led to various response errors9,13,20,21 but not gross articulatory deficits, which instead resulted from DES of structures involved in motor control8,13,20,22 (e.g., the precentral gyrus). Furthermore, perturbation of the cIFG and cMFG delayed inter-speaker timing-consistent with slowed planning-while faster responses could result from stimulation of sites located in other areas. Taken together, our findings suggest that the cIFG and cMFG contain critical preparatory circuits that are relevant for interactive language use.


Assuntos
Fala , Humanos , Masculino , Adulto , Fala/fisiologia , Feminino , Estimulação Elétrica , Córtex Pré-Frontal/fisiologia , Lobo Frontal/fisiologia , Adulto Jovem , Eletrocorticografia , Pessoa de Meia-Idade
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA