Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 1.635
Filter
Add more filters

Publication year range
1.
Int J Yoga Therap ; 34(2024)2024 Apr 01.
Article in English | MEDLINE | ID: mdl-38640400

ABSTRACT

A previous study discovered that two speakers with moderate apraxia of speech increased their sequential motion rates after unilateral forced-nostril breathing (UFNB) practiced as an adjunct to speech-language therapy in an AB repeated-measures design. The current study sought to: (1) delineate possible UFNB plus practice effects from practice effects alone in motor speech skills; (2) examine the relationships between UFNB integrity, participant-reported stress levels, and motor speech performance; and (3) sample a participant-led UFNB training schedule to contribute to the literature's growing understanding of UFNB dosage. A single-subject (n-of-1 trial), ABAB reversal design was used across four motor speech behaviors. A 60-year-old female with chronic, severe apraxia of speech participated. The researchers developed a breathing app to assess UFNB practice integrity and administer the Simple Aphasia Stress Scale after each UFNB session. The participant improved from overall severe to moderate apraxia of speech on the Apraxia Battery for Adults. Visual inspection of graphs confirmed robust motor speech practice effects for all variables. Articulatory-kinematic variables demonstrated sensitivity to the UFNB-plus-practice condition and correlated to stress scale scores but not UFNB integrity scores. The participant achieved 20-minute UFNB sessions 4 times per week. Removal of UFNB during A2 (UFNB withdrawal) and after a 10-day break during B2 (UFNB full dosage) revealed UFNB practice effects on stress scale scores. UFNB with motor speech practice may benefit articulatory-kinematic skills compared to motor speech practice alone. Regular, cumulative UFNB practice appeared to lower self-perceived stress levels. These findings, along with prior work, provide a foundation to further explore yoga breathing and its use with speakers who have apraxia of speech.


Subject(s)
Aphasia , Apraxias , Yoga , Adult , Female , Humans , Middle Aged , Speech , Apraxias/therapy , Respiration , Aphasia/therapy
2.
BMC Health Serv Res ; 24(1): 347, 2024 Mar 16.
Article in English | MEDLINE | ID: mdl-38491356

ABSTRACT

BACKGROUND: Establishing the most important outcomes for school-based speech-language therapy is essential to guide future research and program evaluation for these services. Many health disciplines have developed core outcomes sets (COS) for this purpose. A COS encompasses the most important outcomes for particular health services as identified by appropriate interested parties. These interested parties usually represent health care providers and those with the health condition. In this paper, we report the development of a guiding framework for a COS for speech-language therapy services in schools in a Canadian context. METHODS: Using a group concept mapping method, we identified the outcomes for inclusion in the COS guiding framework through the elicited opinions of key interested parties: speech-language therapists, teachers, and family members of children with speech, language, and communication needs. We extracted 103 statements (potential outcomes) from a previous data set of interview transcripts. We then asked participants to sort the statements into conceptually similar groups, which were aggregated and transformed into a cluster map using multidimensional scaling followed by hierarchical cluster analysis. Participants also rated each statement on 5-point scales for importance and feasibility. We calculated mean ratings for individual statements and for all statements in a cluster, for all participants and for participant groups separately. RESULTS: We identified seven core outcomes for school-based speech-language services in Ontario, Canada. These included: classroom-based services, a holistic approach, support for teachers, care coordination, accessible services, family supports, and student success. All outcomes were rated highly for importance. Feasibility ratings were consistently below importance ratings. All participant groups concurred that a holistic approach was the most important outcome and accessible services was the least feasible outcome to achieve. CONCLUSIONS: The seven outcomes identified in this study are recommended to guide the development of a full COS to direct future research and program evaluation for school-based speech-language services. These outcomes have not been widely included in previous research and should be incorporated into future research alongside specific intervention outcomes. Data for some outcomes may be available from non-traditional sources such as administrative data sets. Consequently, their use for program evaluations should be accompanied by appropriate institutional support to allow speech-language therapists to make meaningful use of appropriate outcomes data.


Subject(s)
Speech Therapy , Speech , Child , Humans , Ontario , Schools , Outcome Assessment, Health Care
3.
PLoS Biol ; 22(3): e3002534, 2024 Mar.
Article in English | MEDLINE | ID: mdl-38466713

ABSTRACT

Selective attention-related top-down modulation plays a significant role in separating relevant speech from irrelevant background speech when vocal attributes separating concurrent speakers are small and continuously evolving. Electrophysiological studies have shown that such top-down modulation enhances neural tracking of attended speech. Yet, the specific cortical regions involved remain unclear due to the limited spatial resolution of most electrophysiological techniques. To overcome such limitations, we collected both electroencephalography (EEG) (high temporal resolution) and functional magnetic resonance imaging (fMRI) (high spatial resolution), while human participants selectively attended to speakers in audiovisual scenes containing overlapping cocktail party speech. To utilise the advantages of the respective techniques, we analysed neural tracking of speech using the EEG data and performed representational dissimilarity-based EEG-fMRI fusion. We observed that attention enhanced neural tracking and modulated EEG correlates throughout the latencies studied. Further, attention-related enhancement of neural tracking fluctuated in predictable temporal profiles. We discuss how such temporal dynamics could arise from a combination of interactions between attention and prediction as well as plastic properties of the auditory cortex. EEG-fMRI fusion revealed attention-related iterative feedforward-feedback loops between hierarchically organised nodes of the ventral auditory object related processing stream. Our findings support models where attention facilitates dynamic neural changes in the auditory cortex, ultimately aiding discrimination of relevant sounds from irrelevant ones while conserving neural resources.


Subject(s)
Auditory Cortex , Speech Perception , Humans , Speech Perception/physiology , Speech , Feedback , Electroencephalography/methods , Auditory Cortex/physiology , Acoustic Stimulation/methods
4.
Commun Biol ; 7(1): 291, 2024 Mar 08.
Article in English | MEDLINE | ID: mdl-38459110

ABSTRACT

When engaged in a conversation, one receives auditory information from the other's speech but also from their own speech. However, this information is processed differently by an effect called Speech-Induced Suppression. Here, we studied brain representation of acoustic properties of speech in natural unscripted dialogues, using electroencephalography (EEG) and high-quality speech recordings from both participants. Using encoding techniques, we were able to reproduce a broad range of previous findings on listening to another's speech, and achieving even better performances when predicting EEG signal in this complex scenario. Furthermore, we found no response when listening to oneself, using different acoustic features (spectrogram, envelope, etc.) and frequency bands, evidencing a strong effect of SIS. The present work shows that this mechanism is present, and even stronger, during natural dialogues. Moreover, the methodology presented here opens the possibility of a deeper understanding of the related mechanisms in a wider range of contexts.


Subject(s)
Electroencephalography , Speech , Humans , Speech/physiology , Acoustic Stimulation/methods , Electroencephalography/methods , Brain , Brain Mapping/methods
5.
PLoS One ; 19(2): e0297826, 2024.
Article in English | MEDLINE | ID: mdl-38330068

ABSTRACT

Perception of sounds and speech involves structures in the auditory brainstem that rapidly process ongoing auditory stimuli. The role of these structures in speech processing can be investigated by measuring their electrical activity using scalp-mounted electrodes. However, typical analysis methods involve averaging neural responses to many short repetitive stimuli that bear little relevance to daily listening environments. Recently, subcortical responses to more ecologically relevant continuous speech were detected using linear encoding models. These methods estimate the temporal response function (TRF), which is a regression model that minimises the error between the measured neural signal and a predictor derived from the stimulus. Using predictors that model the highly non-linear peripheral auditory system may improve linear TRF estimation accuracy and peak detection. Here, we compare predictors from both simple and complex peripheral auditory models for estimating brainstem TRFs on electroencephalography (EEG) data from 24 participants listening to continuous speech. We also investigate the data length required for estimating subcortical TRFs, and find that around 12 minutes of data is sufficient for clear wave V peaks (>3 dB SNR) to be seen in nearly all participants. Interestingly, predictors derived from simple filterbank-based models of the peripheral auditory system yield TRF wave V peak SNRs that are not significantly different from those estimated using a complex model of the auditory nerve, provided that the nonlinear effects of adaptation in the auditory system are appropriately modelled. Crucially, computing predictors from these simpler models is more than 50 times faster compared to the complex model. This work paves the way for efficient modelling and detection of subcortical processing of continuous speech, which may lead to improved diagnosis metrics for hearing impairment and assistive hearing technology.


Subject(s)
Speech Perception , Speech , Humans , Speech Perception/physiology , Hearing/physiology , Brain Stem/physiology , Electroencephalography/methods , Acoustic Stimulation
6.
Sci Rep ; 14(1): 3262, 2024 02 08.
Article in English | MEDLINE | ID: mdl-38332159

ABSTRACT

The McGurk effect refers to an audiovisual speech illusion where the discrepant auditory and visual syllables produce a fused percept between the visual and auditory component. However, little is known about how individual differences contribute to the McGurk effect. Here, we examined whether music training experience-which involves audiovisual integration-can modulate the McGurk effect. Seventy-three participants completed the Goldsmiths Musical Sophistication Index (Gold-MSI) questionnaire to evaluate their music expertise on a continuous scale. Gold-MSI considers participants' daily-life exposure to music learning experiences (formal and informal), instead of merely classifying people into different groups according to how many years they have been trained in music. Participants were instructed to report, via a 3-alternative forced choice task, "what a person said": /Ba/, /Ga/ or /Da/. The experiment consisted of 96 audiovisual congruent trials and 96 audiovisual incongruent (McGurk) trials. We observed no significant correlations between the susceptibility of the McGurk effect and the different subscales of the Gold-MSI (active engagement, perceptual abilities, music training, singing abilities, emotion) or the general musical sophistication composite score. Together, these findings suggest that music training experience does not modulate audiovisual integration in speech as reflected by the McGurk effect.


Subject(s)
Music , Speech Perception , Humans , Visual Perception , Speech , Gold , Auditory Perception , Acoustic Stimulation
7.
J Neurosci ; 44(10)2024 Mar 06.
Article in English | MEDLINE | ID: mdl-38199864

ABSTRACT

During communication in real-life settings, our brain often needs to integrate auditory and visual information and at the same time actively focus on the relevant sources of information, while ignoring interference from irrelevant events. The interaction between integration and attention processes remains poorly understood. Here, we use rapid invisible frequency tagging and magnetoencephalography to investigate how attention affects auditory and visual information processing and integration, during multimodal communication. We presented human participants (male and female) with videos of an actress uttering action verbs (auditory; tagged at 58 Hz) accompanied by two movie clips of hand gestures on both sides of fixation (attended stimulus tagged at 65 Hz; unattended stimulus tagged at 63 Hz). Integration difficulty was manipulated by a lower-order auditory factor (clear/degraded speech) and a higher-order visual semantic factor (matching/mismatching gesture). We observed an enhanced neural response to the attended visual information during degraded speech compared to clear speech. For the unattended information, the neural response to mismatching gestures was enhanced compared to matching gestures. Furthermore, signal power at the intermodulation frequencies of the frequency tags, indexing nonlinear signal interactions, was enhanced in the left frontotemporal and frontal regions. Focusing on the left inferior frontal gyrus, this enhancement was specific for the attended information, for those trials that benefitted from integration with a matching gesture. Together, our results suggest that attention modulates audiovisual processing and interaction, depending on the congruence and quality of the sensory input.


Subject(s)
Brain , Speech Perception , Humans , Male , Female , Brain/physiology , Visual Perception/physiology , Magnetoencephalography , Speech/physiology , Attention/physiology , Speech Perception/physiology , Acoustic Stimulation , Photic Stimulation
8.
Cereb Cortex ; 34(2)2024 01 31.
Article in English | MEDLINE | ID: mdl-38212291

ABSTRACT

Plasticity from auditory experience shapes the brain's encoding and perception of sound. However, whether such long-term plasticity alters the trajectory of short-term plasticity during speech processing has yet to be investigated. Here, we explored the neural mechanisms and interplay between short- and long-term neuroplasticity for rapid auditory perceptual learning of concurrent speech sounds in young, normal-hearing musicians and nonmusicians. Participants learned to identify double-vowel mixtures during ~ 45 min training sessions recorded simultaneously with high-density electroencephalography (EEG). We analyzed frequency-following responses (FFRs) and event-related potentials (ERPs) to investigate neural correlates of learning at subcortical and cortical levels, respectively. Although both groups showed rapid perceptual learning, musicians showed faster behavioral decisions than nonmusicians overall. Learning-related changes were not apparent in brainstem FFRs. However, plasticity was highly evident in cortex, where ERPs revealed unique hemispheric asymmetries between groups suggestive of different neural strategies (musicians: right hemisphere bias; nonmusicians: left hemisphere). Source reconstruction and the early (150-200 ms) time course of these effects localized learning-induced cortical plasticity to auditory-sensory brain areas. Our findings reinforce the domain-general benefits of musicianship but reveal that successful speech sound learning is driven by a critical interplay between long- and short-term mechanisms of auditory plasticity, which first emerge at a cortical level.


Subject(s)
Auditory Cortex , Speech Perception , Humans , Speech , Speech Perception/physiology , Auditory Cortex/physiology , Learning , Electroencephalography , Neuronal Plasticity/physiology , Acoustic Stimulation
9.
Appl Psychophysiol Biofeedback ; 49(1): 71-83, 2024 Mar.
Article in English | MEDLINE | ID: mdl-38165498

ABSTRACT

Biofeedback therapy is mainly based on the analysis of physiological features to improve an individual's affective state. There are insufficient objective indicators to assess symptom improvement after biofeedback. In addition to psychological and physiological features, speech features can precisely convey information about emotions. The use of speech features can improve the objectivity of psychiatric assessments. Therefore, biofeedback based on subjective symptom scales, objective speech, and physiological features to evaluate efficacy provides a new approach for early screening and treatment of emotional problems in college students. A 4-week, randomized, controlled, parallel biofeedback therapy study was conducted with college students with symptoms of anxiety or depression. Speech samples, physiological samples, and clinical symptoms were collected at baseline and at the end of treatment, and the extracted speech features and physiological features were used for between-group comparisons and correlation analyses between the biofeedback and wait-list groups. Based on the speech features with differences between the biofeedback intervention and wait-list groups, an artificial neural network was used to predict the therapeutic effect and response after biofeedback therapy. Through biofeedback therapy, improvements in depression (p = 0.001), anxiety (p = 0.001), insomnia (p = 0.013), and stress (p = 0.004) severity were observed in college-going students (n = 52). The speech and physiological features in the biofeedback group also changed significantly compared to the waitlist group (n = 52) and were related to the change in symptoms. The energy parameters and Mel-Frequency Cepstral Coefficients (MFCC) of speech features can predict whether biofeedback intervention effectively improves anxiety and insomnia symptoms and treatment response. The accuracy of the classification model built using the artificial neural network (ANN) for treatment response and non-response was approximately 60%. The results of this study provide valuable information about biofeedback in improving the mental health of college-going students. The study identified speech features, such as the energy parameters, and MFCC as more accurate and objective indicators for tracking biofeedback therapy response and predicting efficacy. Trial Registration ClinicalTrials.gov ChiCTR2100045542.


Subject(s)
Sleep Initiation and Maintenance Disorders , Speech , Humans , Biofeedback, Psychology/methods , Students/psychology , Biomarkers , Machine Learning
10.
Cortex ; 171: 287-307, 2024 Feb.
Article in English | MEDLINE | ID: mdl-38061210

ABSTRACT

The spectral formant structure and periodicity pitch are the major features that determine the identity of vowels and the characteristics of the speaker. However, very little is known about how the processing of these features in the auditory cortex changes during development. To address this question, we independently manipulated the periodicity and formant structure of vowels while measuring auditory cortex responses using magnetoencephalography (MEG) in children aged 7-12 years and adults. We analyzed the sustained negative shift of source current associated with these vowel properties, which was present in the auditory cortex in both age groups despite differences in the transient components of the auditory response. In adults, the sustained activation associated with formant structure was lateralized to the left hemisphere early in the auditory processing stream requiring neither attention nor semantic mapping. This lateralization was not yet established in children, in whom the right hemisphere contribution to formant processing was strong and decreased during or after puberty. In contrast to the formant structure, periodicity was associated with a greater response in the right hemisphere in both children and adults. These findings suggest that left-lateralization for the automatic processing of vowel formant structure emerges relatively late in ontogenesis and pose a serious challenge to current theories of hemispheric specialization for speech processing.


Subject(s)
Auditory Cortex , Speech Perception , Adult , Humans , Child , Auditory Cortex/physiology , Acoustic Stimulation , Auditory Perception/physiology , Magnetoencephalography , Speech/physiology , Speech Perception/physiology
11.
Clin Child Psychol Psychiatry ; 29(1): 301-311, 2024 Jan.
Article in English | MEDLINE | ID: mdl-37225162

ABSTRACT

Tourette syndrome is characterized by at least two motor tics and one vocal tic, which persist for over a year. Infrequently, tics can manifest as blocking tics in speech when they prevent a person from starting to speak or interrupt their speech flow. Vocal blocking tics (VBTs) resemble stuttering, and they can be difficult to differentiate from each other. A previous report described two patients with severe VBTs who did not benefit from stuttering-therapy-based speech therapy and were treated effectively with cannabis-based medicine. Here, we present the cases of two patients, seven- and nine-year-old boys, who benefited from speech therapy in which stuttering therapy techniques were used. Detailed descriptions of the interventions are included. Further research is needed to test the effectiveness of speech therapy in treating VBTs in a larger group of children with Tourette syndrome.


Subject(s)
Stuttering , Tics , Tourette Syndrome , Male , Child , Humans , Tourette Syndrome/complications , Tourette Syndrome/therapy , Tics/therapy , Tics/etiology , Stuttering/therapy , Stuttering/complications , Speech Therapy , Speech
12.
Neurosurgery ; 94(2): 307-316, 2024 02 01.
Article in English | MEDLINE | ID: mdl-37695053

ABSTRACT

BACKGROUND AND OBJECTIVES: The human thalamus is known, from stimulation studies and functional imaging, to participate in high-level language tasks. The goal of this study is to find whether and how speech features, in particular, vowel phonemes, are encoded in the neuronal activity of the thalamus, and specifically of the left ventralis intermediate nucleus (Vim), during speech production, perception, and imagery. METHODS: In this cross-sectional study, we intraoperatively recorded single neuron activity in the left Vim of eight neurosurgical patients with Parkinson's disease (PD) (n = 4) or essential tremor (n = 4) undergoing implantation of deep brain stimulation (n = 3) or radiofrequency lesioning (n = 5) while patients articulated the five monophthongal vowel sounds. RESULTS: In this article, we report that single neurons in the left Vim encode individual vowel phonemes mainly during speech production but also during perception and imagery. They mainly use one of two encoding schemes: broad or sharp tuning, with a similar percentage of units each. Sinusoidal tuning has been demonstrated in almost half of the broadly tuned units. Patients with PD had a lower percentage of speech-related units in each aspect of speech (production, perception, and imagery), a significantly lower percentage of broadly tuned units, and significantly lower median firing rates during speech production and perception, but significantly higher rates during imagery, than patients with essential tremor. CONCLUSION: The results suggest that the left Vim uses mixed encoding schemes for speech features. Our findings explain, at the single neuron level, why deep brain stimulation and radiofrequency lesioning of the left Vim are likely to cause speech side effects. Moreover, they may indicate that speech-related units in the left Vim of patients with PD may be degraded even in the subclinical phase.


Subject(s)
Deep Brain Stimulation , Essential Tremor , Parkinson Disease , Humans , Parkinson Disease/therapy , Essential Tremor/therapy , Speech , Cross-Sectional Studies , Thalamus , Neurons/physiology , Deep Brain Stimulation/methods
13.
J Behav Ther Exp Psychiatry ; 83: 101940, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38160573

ABSTRACT

BACKGROUND AND OBJECTIVES: It has been proposed that negative mental imagery plays an important role in the persistence of social fears. Experiencing vivid and distressing 'flashforward' images of a potential social catastrophe appears to be of relevance in speech anxiety. To clarify the role of these images, the current experimental study tested if reducing the vividness and distressing properties of recurring negative flashforward images subsequently reduces anxiety and avoidance tendencies regarding a speech. METHODS: Participants were female undergraduates high in speech anxiety (N = 134) who joined our study online. In the experimental condition, we used a visuospatial dual-task to reduce the vividness and distress of flashforward imagery. Primary outcomes were participants' self-reported anxiety and avoidance ratings in anticipation of and during an actual speech. As a secondary outcome, we used observer ratings of participants' anxiety during the speech. RESULTS: Participants reported moderate to high frequency and interference of their vivid and distressing flashforward images in daily life. The dual-task resulted in reductions in image vividness and distress. However, we found no differences between conditions in anxiety and avoidance ratings before and during the speech. LIMITATIONS: The imagery manipulation effect was moderate to small. Moreover, we included a subclinical sample. CONCLUSIONS: Reducing negative flashforward imagery vividness and distress with a visuospatial dual-task did not directly lead to less anxiety and avoidance tendencies related to a later speech. Thus, findings provided no support for the hypothesis that experiencing highly vivid and distressing flashforward images causally contributes to social fears.


Subject(s)
Imagination , Speech , Humans , Female , Male , Anxiety , Imagery, Psychotherapy , Fear
14.
J Assoc Res Otolaryngol ; 24(6): 619-631, 2023 Dec.
Article in English | MEDLINE | ID: mdl-38079021

ABSTRACT

PURPOSE: The role of the medial olivocochlear system in speech perception in noise has been debated over the years, with studies showing mixed results. One possible reason for this could be the dependence of this relationship on the parameters used in assessing the speech perception ability (age, stimulus, and response-related variables). METHODS: The current study assessed the influence of the type of speech stimuli (monosyllables, words, and sentences), the signal-to-noise ratio (+5, 0, -5, and -10 dB), the metric used to quantify the speech perception ability (percent-correct, SNR-50, and slope of the psychometric function) and age (young vs old) on the relationship between medial olivocochlear reflex (quantified by contralateral inhibition of transient evoked otoacoustic emissions) and speech perception in noise. RESULTS: A linear mixed-effects model revealed no significant contributions of the medial olivocochlear reflex to speech perception in noise. CONCLUSION: The results suggest that there was no evidence of any modulatory influence of the indirectly measured medial olivocochlear reflex strength on speech perception in noise.


Subject(s)
Speech Perception , Speech Perception/physiology , Otoacoustic Emissions, Spontaneous/physiology , Speech , Noise , Reflex , Cochlea/physiology , Olivary Nucleus/physiology , Acoustic Stimulation
15.
Article in English | MEDLINE | ID: mdl-38083588

ABSTRACT

Brain-computer interface (BCI) based on speech imagery can decode users' verbal intent and help people with motor disabilities communicate naturally. Functional near-infrared spectroscopy (fNIRS) is a commonly used brain signal acquisition method. Asynchronous BCI can response to control commands at any time, which provides great convenience for users. Task state detection, defined as identifying whether user starts or continues covertly articulating, plays an important role in speech imagery BCIs. To better distinguish task state from idle state during speech imagery, this work used fNIRS signals from different brain regions to study the effects of different brain regions on task state detection accuracy. The imagined tonal syllables included four lexical tones and four vowels in Mandarin Chinese. The brain regions that were measured included Broca's area, Wernicke's area, Superior temporal cortex and Motor cortex. Task state detection accuracies of imagining tonal monosyllables with four different tones were analyzed. The average accuracy of four speech imagery tasks based on the whole brain was 0.67 and it was close to 0.69, which was the average accuracy based on Broca's area. The accuracies of Broca's area and the whole brain were significantly higher than those of other brain regions. The findings of this work demonstrated that using a few channels of Broca's area could result in a similar task state detection accuracy to that using all the channels of the brain. Moreover, it was discovered that speech imagery with tone 2/3 tasks yielded higher task state detection accuracy than speech imagery with other tones.


Subject(s)
Motor Cortex , Speech , Humans , Speech/physiology , Brain/diagnostic imaging , Brain/physiology , Imagery, Psychotherapy , Temporal Lobe , Motor Cortex/physiology
16.
Sci Rep ; 13(1): 22682, 2023 12 19.
Article in English | MEDLINE | ID: mdl-38114579

ABSTRACT

When individuals listen to speech, their neural activity phase-locks to the slow temporal rhythm, which is commonly referred to as "neural tracking". The neural tracking mechanism allows for the detection of an attended sound source in a multi-talker situation by decoding neural signals obtained by electroencephalography (EEG), known as auditory attention decoding (AAD). Neural tracking with AAD can be utilized as an objective measurement tool for diverse clinical contexts, and it has potential to be applied to neuro-steered hearing devices. To effectively utilize this technology, it is essential to enhance the accessibility of EEG experimental setup and analysis. The aim of the study was to develop a cost-efficient neural tracking system and validate the feasibility of neural tracking measurement by conducting an AAD task using an offline and real-time decoder model outside the soundproof environment. We devised a neural tracking system capable of conducting AAD experiments using an OpenBCI and Arduino board. Nine participants were recruited to assess the performance of the AAD using the developed system, which involved presenting competing speech signals in an experiment setting without soundproofing. As a result, the offline decoder model demonstrated an average performance of 90%, and real-time decoder model exhibited a performance of 78%. The present study demonstrates the feasibility of implementing neural tracking and AAD using cost-effective devices in a practical environment.


Subject(s)
Speech Perception , Humans , Acoustic Stimulation , Electroencephalography , Speech , Attention
17.
Brain Lang ; 247: 105359, 2023 12.
Article in English | MEDLINE | ID: mdl-37951157

ABSTRACT

Visual information from a speaker's face enhances auditory neural processing and speech recognition. To determine whether auditory memory can be influenced by visual speech, the degree of auditory neural adaptation of an auditory syllable preceded by an auditory, visual, or audiovisual syllable was examined using EEG. Consistent with previous findings and additional adaptation of auditory neurons tuned to acoustic features, stronger adaptation of N1, P2 and N2 auditory evoked responses was observed when the auditory syllable was preceded by an auditory compared to a visual syllable. However, although stronger than when preceded by a visual syllable, lower adaptation was observed when the auditory syllable was preceded by an audiovisual compared to an auditory syllable. In addition, longer N1 and P2 latencies were then observed. These results further demonstrate that visual speech acts on auditory memory but suggest competing visual influences in the case of audiovisual stimulation.


Subject(s)
Speech Perception , Humans , Speech Perception/physiology , Speech , Electroencephalography , Visual Perception/physiology , Auditory Perception/physiology , Evoked Potentials, Auditory/physiology , Acoustic Stimulation , Photic Stimulation
18.
Proc Natl Acad Sci U S A ; 120(49): e2309166120, 2023 Dec 05.
Article in English | MEDLINE | ID: mdl-38032934

ABSTRACT

Neural speech tracking has advanced our understanding of how our brains rapidly map an acoustic speech signal onto linguistic representations and ultimately meaning. It remains unclear, however, how speech intelligibility is related to the corresponding neural responses. Many studies addressing this question vary the level of intelligibility by manipulating the acoustic waveform, but this makes it difficult to cleanly disentangle the effects of intelligibility from underlying acoustical confounds. Here, using magnetoencephalography recordings, we study neural measures of speech intelligibility by manipulating intelligibility while keeping the acoustics strictly unchanged. Acoustically identical degraded speech stimuli (three-band noise-vocoded, ~20 s duration) are presented twice, but the second presentation is preceded by the original (nondegraded) version of the speech. This intermediate priming, which generates a "pop-out" percept, substantially improves the intelligibility of the second degraded speech passage. We investigate how intelligibility and acoustical structure affect acoustic and linguistic neural representations using multivariate temporal response functions (mTRFs). As expected, behavioral results confirm that perceived speech clarity is improved by priming. mTRFs analysis reveals that auditory (speech envelope and envelope onset) neural representations are not affected by priming but only by the acoustics of the stimuli (bottom-up driven). Critically, our findings suggest that segmentation of sounds into words emerges with better speech intelligibility, and most strongly at the later (~400 ms latency) word processing stage, in prefrontal cortex, in line with engagement of top-down mechanisms associated with priming. Taken together, our results show that word representations may provide some objective measures of speech comprehension.


Subject(s)
Speech Intelligibility , Speech Perception , Speech Intelligibility/physiology , Acoustic Stimulation/methods , Speech/physiology , Noise , Acoustics , Magnetoencephalography/methods , Speech Perception/physiology
19.
Neuroimage ; 282: 120404, 2023 11 15.
Article in English | MEDLINE | ID: mdl-37806465

ABSTRACT

Despite the distortion of speech signals caused by unavoidable noise in daily life, our ability to comprehend speech in noisy environments is relatively stable. However, the neural mechanisms underlying reliable speech-in-noise comprehension remain to be elucidated. The present study investigated the neural tracking of acoustic and semantic speech information during noisy naturalistic speech comprehension. Participants listened to narrative audio recordings mixed with spectrally matched stationary noise at three signal-to-ratio (SNR) levels (no noise, 3 dB, -3 dB), and 60-channel electroencephalography (EEG) signals were recorded. A temporal response function (TRF) method was employed to derive event-related-like responses to the continuous speech stream at both the acoustic and the semantic levels. Whereas the amplitude envelope of the naturalistic speech was taken as the acoustic feature, word entropy and word surprisal were extracted via the natural language processing method as two semantic features. Theta-band frontocentral TRF responses to the acoustic feature were observed at around 400 ms following speech fluctuation onset over all three SNR levels, and the response latencies were more delayed with increasing noise. Delta-band frontal TRF responses to the semantic feature of word entropy were observed at around 200 to 600 ms leading to speech fluctuation onset over all three SNR levels. The response latencies became more leading with increasing noise and decreasing speech comprehension and intelligibility. While the following responses to speech acoustics were consistent with previous studies, our study revealed the robustness of leading responses to speech semantics, which suggests a possible predictive mechanism at the semantic level for maintaining reliable speech comprehension in noisy environments.


Subject(s)
Comprehension , Speech Perception , Humans , Comprehension/physiology , Semantics , Speech/physiology , Speech Perception/physiology , Electroencephalography , Acoustics , Acoustic Stimulation
20.
Am J Case Rep ; 24: e940439, 2023 Oct 03.
Article in English | MEDLINE | ID: mdl-37786244

ABSTRACT

BACKGROUND Brown-Vialetto-Van Laere (BVVL) syndrome is a rare autosomal recessive disorder caused by mutations in intestinal riboflavin transporter genes, resulting in a motor neuron disorder of childhood, which can be associated with sensorineural deafness. This report describes a 4-year-old Polish girl with progressive hearing loss and delayed speech development diagnosed with Brown-Vialetto-Van Laere syndrome who was treated with riboflavin (vitamin B2) and cochlear implants. CASE REPORT The case report concerns a girl from Poland who, at the age of 2 years 10 months, developed progressive atypical neurological symptoms of unknown etiology: ataxia of the upper and lower limbs, gait abnormalities, generalized muscle weakness, visual and hearing problems, and regression of speech development. A karyotype study (whole-exome sequencing) revealed alterations within SLC52A2, leading to the diagnosis of Brown-Vialetto-Van Laere syndrome and initiation of high-dose riboflavin treatment. As a 4-year-old child, she presented to the Institute of Physiology and Pathology of Hearing - World Hearing Center in Poland with progressive hearing loss and speech regression. Hearing tests revealed bilateral profound sensorineural hearing loss with auditory neuropathy. Surgical treatment was applied in the form of bilateral cochlear implantation. CONCLUSIONS This report shows the importance of genetic testing in infants who present with atypical symptoms or signs. In this case, the diagnosis of Brown-Vialetto-Van Laere syndrome resulted in timely correction of the genetic riboflavin (vitamin B2) deficiency and improved hearing following the use of cochlear implants.


Subject(s)
Bulbar Palsy, Progressive , Cochlear Implantation , Cochlear Implants , Deafness , Hearing Loss, Sensorineural , Female , Infant , Humans , Child, Preschool , Speech , Hearing Loss, Sensorineural/etiology , Bulbar Palsy, Progressive/complications , Bulbar Palsy, Progressive/diagnosis , Bulbar Palsy, Progressive/genetics , Riboflavin/therapeutic use , Deafness/complications , Deafness/drug therapy
SELECTION OF CITATIONS
SEARCH DETAIL