Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 28.676
Filter
1.
Sci Rep ; 14(1): 16409, 2024 Jul 16.
Article in English | MEDLINE | ID: mdl-39013983

ABSTRACT

A fundamental aspect of language processing is inferring others' minds from subtle variations in speech. The same word or sentence can often convey different meanings depending on its tempo, timing, and intonation-features often referred to as prosody. Although autistic children and adults are known to experience difficulty in making such inferences, the science remains unclear as to why. We hypothesize that detail-oriented perception in autism may interfere with the inference process if it lacks the adaptivity required to cope with the variability ubiquitous in human speech. Using a novel prosodic continuum that shifts the sentence meaning gradiently from a statement (e.g., "It's raining") to a question (e.g., "It's raining?"), we have investigated the perception and adaptation of receptive prosody in autistic adolescents and two groups of non-autistic controls. Autistic adolescents showed attenuated adaptivity in categorizing prosody, whereas they were equivalent to controls in terms of discrimination accuracy. Combined with recent findings in segmental (e.g., phoneme) recognition, the current results provide the basis for an emerging research framework for attenuated flexibility and reduced influence of contextual feedback as a possible source of deficits that hinder linguistic and social communication in autism.


Subject(s)
Autistic Disorder , Speech Perception , Humans , Adolescent , Male , Female , Speech Perception/physiology , Autistic Disorder/physiopathology , Autistic Disorder/psychology , Language , Child , Speech/physiology
2.
PLoS One ; 19(7): e0305657, 2024.
Article in English | MEDLINE | ID: mdl-39018339

ABSTRACT

Technological developments over the past few decades have changed the way people communicate, with platforms like social media and blogs becoming vital channels for international conversation. Even though hate speech is vigorously suppressed on social media, it is still a concern that needs to be constantly recognized and observed. The Arabic language poses particular difficulties in the detection of hate speech, despite the considerable efforts made in this area for English-language social media content. Arabic calls for particular consideration when it comes to hate speech detection because of its many dialects and linguistic nuances. Another degree of complication is added by the widespread practice of "code-mixing," in which users merge various languages smoothly. Recognizing this research vacuum, the study aims to close it by examining how well machine learning models containing variation features can detect hate speech, especially when it comes to Arabic tweets featuring code-mixing. Therefore, the objective of this study is to assess and compare the effectiveness of different features and machine learning models for hate speech detection on Arabic hate speech and code-mixing hate speech datasets. To achieve the objectives, the methodology used includes data collection, data pre-processing, feature extraction, the construction of classification models, and the evaluation of the constructed classification models. The findings from the analysis revealed that the TF-IDF feature, when employed with the SGD model, attained the highest accuracy, reaching 98.21%. Subsequently, these results were contrasted with outcomes from three existing studies, and the proposed method outperformed them, underscoring the significance of the proposed method. Consequently, our study carries practical implications and serves as a foundational exploration in the realm of automated hate speech detection in text.


Subject(s)
Language , Machine Learning , Social Media , Humans , Speech/physiology
3.
Sci Rep ; 14(1): 16603, 2024 Jul 18.
Article in English | MEDLINE | ID: mdl-39025957

ABSTRACT

Electrophysiological brain activity has been shown to synchronize with the quasi-regular repetition of grammatical phrases in connected speech-so-called phrase-rate neural tracking. Current debate centers around whether this phenomenon is best explained in terms of the syntactic properties of phrases or in terms of syntax-external information, such as the sequential repetition of parts of speech. As these two factors were confounded in previous studies, much of the literature is compatible with both accounts. Here, we used electroencephalography (EEG) to determine if and when the brain is sensitive to both types of information. Twenty native speakers of Mandarin Chinese listened to isochronously presented streams of monosyllabic words, which contained either grammatical two-word phrases (e.g., catch fish, sell house) or non-grammatical word combinations (e.g., full lend, bread far). Within the grammatical conditions, we varied two structural factors: the position of the head of each phrase and the type of attachment. Within the non-grammatical conditions, we varied the consistency with which parts of speech were repeated. Tracking was quantified through evoked power and inter-trial phase coherence, both derived from the frequency-domain representation of EEG responses. As expected, neural tracking at the phrase rate was stronger in grammatical sequences than in non-grammatical sequences without syntactic structure. Moreover, it was modulated by both attachment type and head position, revealing the structure-sensitivity of phrase-rate tracking. We additionally found that the brain tracks the repetition of parts of speech in non-grammatical sequences. These data provide an integrative perspective on the current debate about neural tracking effects, revealing that the brain utilizes regularities computed over multiple levels of linguistic representation in guiding rhythmic computation.


Subject(s)
Brain , Electroencephalography , Humans , Male , Female , Adult , Brain/physiology , Young Adult , Language , Speech Perception/physiology , Speech/physiology
4.
PLoS One ; 19(7): e0306272, 2024.
Article in English | MEDLINE | ID: mdl-39028710

ABSTRACT

Abnormal speech prosody has been widely reported in individuals with autism. Many studies on children and adults with autism spectrum disorder speaking a non-tonal language showed deficits in using prosodic cues to mark focus. However, focus marking by autistic children speaking a tonal language is rarely examined. Cantonese-speaking children may face additional difficulties because tonal languages require them to use prosodic cues to achieve multiple functions simultaneously such as lexical contrasting and focus marking. This study bridges this research gap by acoustically evaluating the use of Cantonese speech prosody to mark information structure by Cantonese-speaking children with and without autism spectrum disorder. We designed speech production tasks to elicit natural broad and narrow focus production among these children in sentences with different tone combinations. Acoustic correlates of prosodic focus marking like f0, duration and intensity of each syllable were analyzed to examine the effect of participant group, focus condition and lexical tones. Our results showed differences in focus marking patterns between Cantonese-speaking children with and without autism spectrum disorder. The autistic children not only showed insufficient on-focus expansion in terms of f0 range and duration when marking focus, but also produced less distinctive tone shapes in general. There was no evidence that the prosodic complexity (i.e. sentences with single tones or combinations of tones) significantly affected focus marking in these autistic children and their typically-developing (TD) peers.


Subject(s)
Autism Spectrum Disorder , Language , Humans , Autism Spectrum Disorder/physiopathology , Autism Spectrum Disorder/psychology , Male , Female , Child , Speech Acoustics , Child, Preschool , Speech/physiology
5.
J Robot Surg ; 18(1): 287, 2024 Jul 18.
Article in English | MEDLINE | ID: mdl-39026112

ABSTRACT

Transoral robotic surgery (TORS) has been introduced to head and neck surgery as a minimally invasive techqniques to improve the functional outcomes of patients. Compare the functional outcome for swallowing and speech in each site of TORS within the head and neck. Retrospective cohort study for patients who underwent TORS within the head and neck unit. Patients were assessed at four different time points (one day, one month, six months and twelve months, respectively) with bedside/office testing. Methods of testing for swallowing assessment were by the International Dysphagia Diet Standardization Initiative (IDDSI), and speech assessments were carried out using the Understandability of Speech score (USS). Outcomes were compared to patient-specific pre-treatment baseline levels. 68 patients were included. 75% and 40% of the patients resumed normal fluid intake and normal diet immediately after surgery. 8.8% required a temporary feeding tube, with 1% required gastrostomy. There was a steep improvement in diet between 3 and 6 months. Fluid and diet consistency dropped significantly following the majority of transoral robotic surgery with more noticeable diet changes. Early deterioration in diet is temporary and manageable with a modified diet. Rapid recovery of swallowing is achieved before the first year. There is no long-term effect on speech.


Subject(s)
Deglutition Disorders , Deglutition , Robotic Surgical Procedures , Speech , Humans , Robotic Surgical Procedures/methods , Deglutition/physiology , Male , Female , Retrospective Studies , Speech/physiology , Middle Aged , Aged , Deglutition Disorders/etiology , Treatment Outcome , Mouth , Adult , Head and Neck Neoplasms/surgery , Aged, 80 and over
6.
Sci Data ; 11(1): 746, 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38982093

ABSTRACT

Many research articles have explored the impact of surgical interventions on voice and speech evaluations, but advances are limited by the lack of publicly accessible datasets. To address this, a comprehensive corpus of 107 Spanish Castilian speakers was recorded, including control speakers and patients who underwent upper airway surgeries such as Tonsillectomy, Functional Endoscopic Sinus Surgery, and Septoplasty. The dataset contains 3,800 audio files, averaging 35.51 ± 5.91 recordings per patient. This resource enables systematic investigation of the effects of upper respiratory tract surgery on voice and speech. Previous studies using this corpus have shown no relevant changes in key acoustic parameters for sustained vowel phonation, consistent with initial hypotheses. However, the analysis of speech recordings, particularly nasalised segments, remains open for further research. Additionally, this dataset facilitates the study of the impact of upper airway surgery on speaker recognition and identification methods, and testing of anti-spoofing methodologies for improved robustness.


Subject(s)
Speech , Voice , Humans , Postoperative Period , Tonsillectomy , Male , Female , Preoperative Period , Adult
7.
Sci Rep ; 14(1): 15787, 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38982177

ABSTRACT

Diagnostic tests for Parkinsonism based on speech samples have shown promising results. Although abnormal auditory feedback integration during speech production and impaired rhythmic organization of speech are known in Parkinsonism, these aspects have not been incorporated into diagnostic tests. This study aimed to identify Parkinsonism using a novel speech behavioral test that involved rhythmically repeating syllables under different auditory feedback conditions. The study included 30 individuals with Parkinson's disease (PD) and 30 healthy subjects. Participants were asked to rhythmically repeat the PA-TA-KA syllable sequence, both whispering and speaking aloud under various listening conditions. The results showed that individuals with PD had difficulties in whispering and articulating under altered auditory feedback conditions, exhibited delayed speech onset, and demonstrated inconsistent rhythmic structure across trials compared to controls. These parameters were then fed into a supervised machine-learning algorithm to differentiate between the two groups. The algorithm achieved an accuracy of 85.4%, a sensitivity of 86.5%, and a specificity of 84.3%. This pilot study highlights the potential of the proposed behavioral paradigm as an objective and accessible (both in cost and time) test for identifying individuals with Parkinson's disease.


Subject(s)
Feedback, Sensory , Parkinson Disease , Speech , Humans , Female , Male , Aged , Parkinson Disease/physiopathology , Parkinson Disease/diagnosis , Middle Aged , Speech/physiology , Feedback, Sensory/physiology , Pilot Projects , Parkinsonian Disorders/physiopathology , Case-Control Studies
8.
Dental Press J Orthod ; 29(3): e2423277, 2024.
Article in English | MEDLINE | ID: mdl-38985077

ABSTRACT

OBJECTIVE: This study aimed to compare the influence of four different maxillary removable orthodontic retainers on speech. MATERIAL AND METHODS: Eligibility criteria for sample selection were: 20-40-year subjects with acceptable occlusion, native speakers of Portuguese. The volunteers (n=21) were divided in four groups randomized with a 1:1:1:1 allocation ratio. The four groups used, in random order, the four types of retainers full-time for 21 days each, with a washout period of 7-days. The removable maxillary retainers were: conventional wraparound, wraparound with an anterior hole, U-shaped wraparound, and thermoplastic retainer. Three volunteers were excluded. The final sample comprised 18 subjects (11 male; 7 female) with mean age of 27.08 years (SD=4.65). The speech evaluation was performed in vocal excerpts recordings made before, immediately after, and 21 days after the installation of each retainer, with auditory-perceptual and acoustic analysis of formant frequencies F1 and F2 of the vowels. Repeated measures ANOVA and Friedman with Tukey tests were used for statistical comparison. RESULTS: Speech changes increased immediately after conventional wraparound and thermoplastic retainer installation, and reduced after 21 days, but not to normal levels. However, this increase was statistically significant only for the wraparound with anterior hole and the thermoplastic retainer. Formant frequencies of vowels were altered at initial time, and the changes remained in conventional, U-shaped and thermoplastic appliances after three weeks. CONCLUSIONS: The thermoplastic retainer was more harmful to the speech than wraparound appliances. The conventional and U-shaped retainers interfered less in speech. The three-week period was not sufficient for speech adaptation.


Subject(s)
Cross-Over Studies , Orthodontic Retainers , Humans , Female , Male , Adult , Orthodontic Appliance Design , Young Adult , Speech/physiology
9.
Sensors (Basel) ; 24(13)2024 Jun 25.
Article in English | MEDLINE | ID: mdl-39000889

ABSTRACT

Emotions in speech are expressed in various ways, and the speech emotion recognition (SER) model may perform poorly on unseen corpora that contain different emotional factors from those expressed in training databases. To construct an SER model robust to unseen corpora, regularization approaches or metric losses have been studied. In this paper, we propose an SER method that incorporates relative difficulty and labeling reliability of each training sample. Inspired by the Proxy-Anchor loss, we propose a novel loss function which gives higher gradients to the samples for which the emotion labels are more difficult to estimate among those in the given minibatch. Since the annotators may label the emotion based on the emotional expression which resides in the conversational context or other modality but is not apparent in the given speech utterance, some of the emotional labels may not be reliable and these unreliable labels may affect the proposed loss function more severely. In this regard, we propose to apply label smoothing for the samples misclassified by a pre-trained SER model. Experimental results showed that the performance of the SER on unseen corpora was improved by adopting the proposed loss function with label smoothing on the misclassified data.


Subject(s)
Emotions , Speech , Humans , Emotions/physiology , Speech/physiology , Algorithms , Reproducibility of Results , Pattern Recognition, Automated/methods , Databases, Factual
10.
Commun Biol ; 7(1): 818, 2024 Jul 05.
Article in English | MEDLINE | ID: mdl-38969758

ABSTRACT

Speech brain-computer interfaces aim to support communication-impaired patients by translating neural signals into speech. While impressive progress was achieved in decoding performed, perceived and attempted speech, imagined speech remains elusive, mainly due to the absence of behavioral output. Nevertheless, imagined speech is advantageous since it does not depend on any articulator movements that might become impaired or even lost throughout the stages of a neurodegenerative disease. In this study, we analyzed electrocortigraphy data recorded from 16 participants in response to 3 speech modes: performed, perceived (listening), and imagined speech. We used a linear model to detect speech events and examined the contributions of each frequency band, from delta to high gamma, given the speech mode and electrode location. For imagined speech detection, we observed a strong contribution of gamma bands in the motor cortex, whereas lower frequencies were more prominent in the temporal lobe, in particular of the left hemisphere. Based on the similarities in frequency patterns, we were able to transfer models between speech modes and participants with similar electrode locations.


Subject(s)
Brain-Computer Interfaces , Electrocorticography , Imagination , Speech , Humans , Electrocorticography/methods , Speech/physiology , Male , Female , Adult , Imagination/physiology , Young Adult , Motor Cortex/physiology
11.
Hum Brain Mapp ; 45(11): e26797, 2024 Aug 01.
Article in English | MEDLINE | ID: mdl-39041175

ABSTRACT

Speech comprehension is crucial for human social interaction, relying on the integration of auditory and visual cues across various levels of representation. While research has extensively studied multisensory integration (MSI) using idealised, well-controlled stimuli, there is a need to understand this process in response to complex, naturalistic stimuli encountered in everyday life. This study investigated behavioural and neural MSI in neurotypical adults experiencing audio-visual speech within a naturalistic, social context. Our novel paradigm incorporated a broader social situational context, complete words, and speech-supporting iconic gestures, allowing for context-based pragmatics and semantic priors. We investigated MSI in the presence of unimodal (auditory or visual) or complementary, bimodal speech signals. During audio-visual speech trials, compared to unimodal trials, participants more accurately recognised spoken words and showed a more pronounced suppression of alpha power-an indicator of heightened integration load. Importantly, on the neural level, these effects surpassed mere summation of unimodal responses, suggesting non-linear MSI mechanisms. Overall, our findings demonstrate that typically developing adults integrate audio-visual speech and gesture information to facilitate speech comprehension in noisy environments, highlighting the importance of studying MSI in ecologically valid contexts.


Subject(s)
Gestures , Speech Perception , Humans , Female , Male , Speech Perception/physiology , Young Adult , Adult , Visual Perception/physiology , Electroencephalography , Comprehension/physiology , Acoustic Stimulation , Speech/physiology , Brain/physiology , Photic Stimulation/methods
12.
Elife ; 132024 Jul 22.
Article in English | MEDLINE | ID: mdl-39038076

ABSTRACT

To what extent does speech and music processing rely on domain-specific and domain-general neural networks? Using whole-brain intracranial EEG recordings in 18 epilepsy patients listening to natural, continuous speech or music, we investigated the presence of frequency-specific and network-level brain activity. We combined it with a statistical approach in which a clear operational distinction is made between shared, preferred, and domain-selective neural responses. We show that the majority of focal and network-level neural activity is shared between speech and music processing. Our data also reveal an absence of anatomical regional selectivity. Instead, domain-selective neural responses are restricted to distributed and frequency-specific coherent oscillations, typical of spectral fingerprints. Our work highlights the importance of considering natural stimuli and brain dynamics in their full complexity to map cognitive and brain functions.


Subject(s)
Music , Humans , Male , Female , Adult , Nerve Net/physiology , Speech/physiology , Auditory Perception/physiology , Epilepsy/physiopathology , Young Adult , Electroencephalography , Cerebral Cortex/physiology , Electrocorticography , Speech Perception/physiology , Middle Aged , Brain Mapping
14.
PLoS One ; 19(7): e0301692, 2024.
Article in English | MEDLINE | ID: mdl-39012881

ABSTRACT

Speech enhancement is crucial both for human and machine listening applications. Over the last decade, the use of deep learning for speech enhancement has resulted in tremendous improvement over the classical signal processing and machine learning methods. However, training a deep neural network is not only time-consuming; it also requires extensive computational resources and a large training dataset. Transfer learning, i.e. using a pretrained network for a new task, comes to the rescue by reducing the amount of training time, computational resources, and the required dataset, but the network still needs to be fine-tuned for the new task. This paper presents a novel method of speech denoising and dereverberation (SD&D) on an end-to-end frozen binaural anechoic speech separation network. The frozen network requires neither any architectural change nor any fine-tuning for the new task, as is usually required for transfer learning. The interaural cues of a source placed inside noisy and echoic surroundings are given as input to this pretrained network to extract the target speech from noise and reverberation. Although the pretrained model used in this paper has never seen noisy reverberant conditions during its training, it performs satisfactorily for zero-shot testing (ZST) under these conditions. It is because the pretrained model used here has been trained on the direct-path interaural cues of an active source and so it can recognize them even in the presence of echoes and noise. ZST on the same dataset on which the pretrained network was trained (homo-corpus) for the unseen class of interference, has shown considerable improvement over the weighted prediction error (WPE) algorithm in terms of four objective speech quality and intelligibility metrics. Also, the proposed model offers similar performance provided by a deep learning SD&D algorithm for this dataset under varying conditions of noise and reverberations. Similarly, ZST on a different dataset has provided an improvement in intelligibility and almost equivalent quality as provided by the WPE algorithm.


Subject(s)
Noise , Humans , Speech , Deep Learning , Signal-To-Noise Ratio , Neural Networks, Computer , Speech Perception/physiology , Algorithms , Signal Processing, Computer-Assisted
15.
Sci Rep ; 14(1): 17135, 2024 Jul 25.
Article in English | MEDLINE | ID: mdl-39054330

ABSTRACT

Nonhuman great apes have been claimed to be unable to learn human words due to a lack of the necessary neural circuitry. We recovered original footage of two enculturated chimpanzees uttering the word "mama" and subjected recordings to phonetic analysis. Our analyses demonstrate that chimpanzees are capable of syllabic production, achieving consonant-to-vowel phonetic contrasts via the simultaneous recruitment and coupling of voice, jaw and lips. In an online experiment, human listeners naive to the recordings' origins reliably perceived chimpanzee utterances as syllabic utterances, primarily as "ma-ma", among foil syllables. Our findings demonstrate that in the absence of direct data-driven examination, great ape vocal production capacities have been underestimated. Chimpanzees possess the neural building blocks necessary for speech.


Subject(s)
Pan troglodytes , Speech , Vocalization, Animal , Pan troglodytes/physiology , Animals , Vocalization, Animal/physiology , Speech/physiology , Humans , Male , Female , Phonetics
16.
Sci Rep ; 14(1): 17110, 2024 Jul 24.
Article in English | MEDLINE | ID: mdl-39048617

ABSTRACT

Research suggests that interlocutors manage the timing demands of conversation by preparing what they want to say early. In three experiments, we used a verbal question-answering task to investigate what aspects of their response speakers prepare early. In all three experiments, participants answered more quickly when the critical content (here, barks) necessary for answer preparation occurred early (e.g., Which animal barks and is also a common household pet?) rather than late (e.g., Which animal is a common household pet and also barks?). In the individual experiments, we found no convincing evidence that participants were slower to produce longer answers, consisting of multiple words, than shorter answers, consisting of a single word. There was also no interaction between these two factors. A combined analysis of the first two experiments confirmed this lack of interaction, and demonstrated that participants were faster to answer questions when the critical content was available early rather than late and when the answer was short rather than long. These findings provide tentative evidence for an account in which interlocutors prepare the content of their answer as soon as they can, but sometimes do not prepare its length (and thus form) until they are ready to speak.


Subject(s)
Communication , Humans , Female , Male , Adult , Young Adult , Reaction Time/physiology , Speech
17.
Cogn Sci ; 48(7): e13478, 2024 Jul.
Article in English | MEDLINE | ID: mdl-38980972

ABSTRACT

How do cognitive pressures shape the lexicons of natural languages? Here, we reframe George Kingsley Zipf's proposed "law of abbreviation" within a more general framework that relates it to cognitive pressures that affect speakers and listeners. In this new framework, speakers' drive to reduce effort (Zipf's proposal) is counteracted by the need for low-frequency words to have word forms that are sufficiently distinctive to allow for accurate recognition by listeners. To support this framework, we replicate and extend recent work using the prevalence of subword phonemic sequences (phonotactic probability) to measure speakers' production effort in place of Zipf's measure of length. Across languages and corpora, phonotactic probability is more strongly correlated with word frequency than word length. We also show this measure of ease of speech production (phonotactic probability) is strongly correlated with a measure of perceptual difficulty that indexes the degree of competition from alternative interpretations in word recognition. This is consistent with the claim that there must be trade-offs between these two factors, and is inconsistent with a recent proposal that phonotactic probability facilitates both perception and production. To our knowledge, this is the first work to offer an explanation why long, phonotactically improbable word forms remain in the lexicons of natural languages.


Subject(s)
Language , Phonetics , Recognition, Psychology , Speech Perception , Humans , Speech
18.
Sci Rep ; 14(1): 15611, 2024 Jul 06.
Article in English | MEDLINE | ID: mdl-38971806

ABSTRACT

This study compares how English-speaking adults and children from the United States adapt their speech when talking to a real person and a smart speaker (Amazon Alexa) in a psycholinguistic experiment. Overall, participants produced more effortful speech when talking to a device (longer duration and higher pitch). These differences also varied by age: children produced even higher pitch in device-directed speech, suggesting a stronger expectation to be misunderstood by the system. In support of this, we see that after a staged recognition error by the device, children increased pitch even more. Furthermore, both adults and children displayed the same degree of variation in their responses for whether "Alexa seems like a real person or not", further indicating that children's conceptualization of the system's competence shaped their register adjustments, rather than an increased anthropomorphism response. This work speaks to models on the mechanisms underlying speech production, and human-computer interaction frameworks, providing support for routinized theories of spoken interaction with technology.


Subject(s)
Speech , Humans , Adult , Child , Male , Female , Speech/physiology , Young Adult , Adolescent , Psycholinguistics
19.
J Neural Eng ; 21(4)2024 Jul 12.
Article in English | MEDLINE | ID: mdl-38925110

ABSTRACT

Objective.Speech brain-computer interfaces (BCIs) have the potential to augment communication in individuals with impaired speech due to muscle weakness, for example in amyotrophic lateral sclerosis (ALS) and other neurological disorders. However, to achieve long-term, reliable use of a speech BCI, it is essential for speech-related neural signal changes to be stable over long periods of time. Here we study, for the first time, the stability of speech-related electrocorticographic (ECoG) signals recorded from a chronically implanted ECoG BCI over a 12 month period.Approach.ECoG signals were recorded by an ECoG array implanted over the ventral sensorimotor cortex in a clinical trial participant with ALS. Because ECoG-based speech decoding has most often relied on broadband high gamma (HG) signal changes relative to baseline (non-speech) conditions, we studied longitudinal changes of HG band power at baseline and during speech, and we compared these with residual high frequency noise levels at baseline. Stability was further assessed by longitudinal measurements of signal-to-noise ratio, activation ratio, and peak speech-related HG response magnitude (HG response peaks). Lastly, we analyzed the stability of the event-related HG power changes (HG responses) for individual syllables at each electrode.Main Results.We found that speech-related ECoG signal responses were stable over a range of syllables activating different articulators for the first year after implantation.Significance.Together, our results indicate that ECoG can be a stable recording modality for long-term speech BCI systems for those living with severe paralysis.Clinical Trial Information.ClinicalTrials.gov, registration number NCT03567213.


Subject(s)
Amyotrophic Lateral Sclerosis , Brain-Computer Interfaces , Electrocorticography , Speech , Humans , Amyotrophic Lateral Sclerosis/physiopathology , Longitudinal Studies , Electrocorticography/methods , Speech/physiology , Male , Gamma Rhythm/physiology , Middle Aged , Female , Electrodes, Implanted
20.
J Speech Lang Hear Res ; 67(7): 2333-2342, 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38875483

ABSTRACT

PURPOSE: This study explored the use of an automated language analysis tool, FLUCALC, for measuring fluency in aphasia. The purpose was to determine whether CLAN's FLUCALC command could produce efficient, objective outcome measures for salient aspects of fluency in aphasia. METHOD: The FLUCALC command was used on CHAT transcripts of Cinderella stories from people with aphasia (PWA; n = 281) and controls (n = 257) in the AphasiaBank database. RESULTS: PWA produced significantly fewer total words, fewer words per minute, more pausing, more repetitions, more revisions, and more phonological fragments than controls, with only one exception: The Wernicke's group was similar to the control group in percentage of filled pauses. Individuals with Broca's aphasia had significantly longer inter-utterance pauses and fewer total words than all other aphasia groups. Both the Broca's and conduction aphasia groups had higher percentages of phrase repetitions than the NABW (NotAphasicByWAB) group. The conduction aphasia group also had a higher percentage of phrase revisions than the NABW and the anomic aphasia groups. Principal components analysis revealed two principal components that accounted for around 60% of the variance and related to quantity of output, rate of speech, and quality of output. The Gaussian mixture models showed that the participants clustered in three groups, which corresponded predominantly to the controls, the nonfluent aphasia group, and the remaining aphasia groups (all classically fluent aphasia types). CONCLUSIONS: FLUCALC is an efficient way to measure objective fluency behaviors in language samples in aphasia. Automated analyses of objective fluency behaviors on large samples of adults with and without aphasia can produce measures that can be used by researchers and clinicians to better understand and track salient aspects of fluency in aphasia. SUPPLEMENTAL MATERIAL: https://doi.org/10.23641/asha.25979863.


Subject(s)
Aphasia , Humans , Female , Male , Middle Aged , Aged , Aphasia/psychology , Adult , Language Tests , Speech/physiology , Speech Production Measurement/methods , Case-Control Studies , Aged, 80 and over
SELECTION OF CITATIONS
SEARCH DETAIL