Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 144
Filter
1.
Lang Speech ; : 238309241254350, 2024 Jun 10.
Article in English | MEDLINE | ID: mdl-38853599

ABSTRACT

Previous research has shown that it is difficult for English speakers to distinguish the front rounded vowels /y/ and /ø/ from the back rounded vowels /u/ and /o/. In this study, we examine the effect of noise on this perceptual difficulty. In an Oddity Discrimination Task, English speakers without any knowledge of German were asked to discriminate between German-sounding pseudowords varying in the vowel both in quiet and in white noise at two signal-to-noise ratios (8 and 0 dB). In test trials, vowels of the same height were contrasted with each other, whereas a contrast with /a/ served as a control trial. Results revealed that a contrast with /a/ remained stable in every listening condition for both high and mid vowels. When contrasting vowels of the same height, however, there was a perceptual shift along the F2 dimension as the noise level increased. Although the /ø/-/o/ and particularly /y/-/u/ contrasts were the most difficult in quiet, accuracy on /i/-/y/ and /e/-/ø/ trials decreased immensely when the speech signal was masked. The German control group showed the same pattern, albeit less severe than the non-native group, suggesting that even in low-level tasks with pseudowords, there is a native advantage in speech perception in noise.

2.
Front Psychol ; 15: 1414363, 2024.
Article in English | MEDLINE | ID: mdl-38711754

ABSTRACT

[This corrects the article DOI: 10.3389/fpsyg.2024.1333112.].

3.
Cogn Neuropsychol ; : 1-19, 2024 May 22.
Article in English | MEDLINE | ID: mdl-38778635

ABSTRACT

How does cognitive inhibition influence speaking? The Stroop effect is a classic demonstration of the interference between reading and color naming. We used a novel variant of the Stroop task to measure whether this interference impacts not only the response speed, but also the acoustic properties of speech. Speakers named the color of words in three categories: congruent (e.g., red written in red), color-incongruent (e.g., green written in red), and vowel-incongruent - those with partial phonological overlap with their color (e.g., rid written in red, grain in green, and blow in blue). Our primary aim was to identify any effect of the distractor vowel on the acoustics of the target vowel. Participants were no slower to respond on vowel-incongruent trials, but formant trajectories tended to show a bias away from the distractor vowel, consistent with a phenomenon of acoustic inhibition that increases contrast between confusable alternatives.

4.
Clin Linguist Phon ; : 1-19, 2024 Apr 16.
Article in English | MEDLINE | ID: mdl-38626320

ABSTRACT

Learning vowel transcription skills is crucial to function as a Speech and Language Therapist (SLT). However, vowel transcription is commonly regarded as particularly difficult and therefore often avoided. Despite the importance of accurate transcriptions, little is known about all the factors that influence the process of learning vowel transcription, which usually includes the learning of the Cardinal Vowel (CV) system. There are only a few studies that investigate how CVs are learnt and what factors lead to successful learning. The current study reports students' perceived difficulty of producing and transcribing CVs as a first step to identify how perceived difficulty affects phonetic learning. Perceived difficulty ratings for the production and transcription of 12 CVs collected from 155 students studying towards a qualification as an SLT were analysed. The results show that the classificatory features correlate with the perceived task difficulty of production and transcription. Implications for teaching are outlined.

5.
J Neural Eng ; 21(3)2024 May 09.
Article in English | MEDLINE | ID: mdl-38648783

ABSTRACT

Objective. Our goal is to decode firing patterns of single neurons in the left ventralis intermediate nucleus (Vim) of the thalamus, related to speech production, perception, and imagery. For realistic speech brain-machine interfaces (BMIs), we aim to characterize the amount of thalamic neurons necessary for high accuracy decoding.Approach. We intraoperatively recorded single neuron activity in the left Vim of eight neurosurgical patients undergoing implantation of deep brain stimulator or RF lesioning during production, perception and imagery of the five monophthongal vowel sounds. We utilized the Spade decoder, a machine learning algorithm that dynamically learns specific features of firing patterns and is based on sparse decomposition of the high dimensional feature space.Main results. Spade outperformed all algorithms compared with, for all three aspects of speech: production, perception and imagery, and obtained accuracies of 100%, 96%, and 92%, respectively (chance level: 20%) based on pooling together neurons across all patients. The accuracy was logarithmic in the amount of neurons for all three aspects of speech. Regardless of the amount of units employed, production gained highest accuracies, whereas perception and imagery equated with each other.Significance. Our research renders single neuron activity in the left Vim a promising source of inputs to BMIs for restoration of speech faculties for locked-in patients or patients with anarthria or dysarthria to allow them to communicate again. Our characterization of how many neurons are necessary to achieve a certain decoding accuracy is of utmost importance for planning BMI implantation.


Subject(s)
Brain-Computer Interfaces , Machine Learning , Neurons , Speech , Thalamus , Humans , Neurons/physiology , Male , Female , Middle Aged , Speech/physiology , Adult , Thalamus/physiology , Deep Brain Stimulation/methods , Aged , Speech Perception/physiology
6.
Front Psychol ; 15: 1333112, 2024.
Article in English | MEDLINE | ID: mdl-38524296

ABSTRACT

Purpose: The main purpose of this study was to investigate to what extent the L2 Arabic learners' reading process is affected by the incomplete representation of speech (the absence of short vowels and diacritics) while reading ambiguous sentences (garden path sentences). Method: With a self-paced reading software program, 41 non-native male students, aged from 22 to 26, enrolled in King Saud University, participated in reading 44 sentences (followed by reading comprehension questions) representing three reading conditions, plain, vowelized-discretized, and wrongly-vowelized. Results: For the reading times data, the analysis revealed that the GP structure had a significant effect on the reading processes of L2 Arabic learners; it took them longer to read the GP sentences than their non-GP counterparts. For the reading comprehension, the analysis did not reveal any significant differences between the means for the percentages of correct responses. For the comparison between the three reading conditions, a significant difference was found: it took the participants on average less time to read the GP sentences when presented plain, and more time with the incorrect representation. However, their reading comprehension was not affected. Conclusion: In addition to the good-enough model and the nature of Arabic morphology, the reading experience, is a good candidate to start with as an important factor in the interpretation of the ineffectiveness of the GP structure on the reading comprehension process of Arabic readers, in which the segregability of Arabic writing system prepare the readers to emphasize some sensory inputs and ignore others based on their past reading experience.

7.
Lang Speech ; : 238309231223909, 2024 Feb 05.
Article in English | MEDLINE | ID: mdl-38312096

ABSTRACT

To date, research on wordform learning biases has mostly focused on language-dependent factors, such as the phonotactics and neighborhood density of the language(s) known by the learner. Domain-general biases, by contrast, have received little attention. In this study, we focus on one such bias-an advantage for string-internal repetitions-and examine its effects on wordform learning. Importantly, we consider whether any type of segmental repetition is equally beneficial for word recall, or whether learning is favored more or only by repeated consonants, in line with previous research indicating that consonants play a larger role than vowels in lexical processing. In Experiment 1, adult English speakers learned artificial consonant-vowel-consonant-vowel words containing either a repeated consonant (e.g., /sesu/, "c-rep"), a repeated vowel (e.g., /sepe/, "v-rep"), or dissimilar consonants and vowels (e.g., /sepu/, "no-rep"). Recall results showed no advantage for v-reps but higher accuracy for c-reps compared with no-reps. In Experiment 2, participants performed a label preference task with the same stimuli. The results showed dispreference for both c-reps and v-reps relative to no-reps, indicating that the results of Experiment 1 are independent of wordlikeness effects. These outcomes reveal that there is a form-learning bias for words with identical consonants but not for words with identical vowels, suggesting that a domain-general advantage for repetitions within strings is modulated by a language-specific processing bias for consonants.

8.
Q J Exp Psychol (Hove) ; : 17470218241229721, 2024 Feb 16.
Article in English | MEDLINE | ID: mdl-38262925

ABSTRACT

The contribution of consonants and vowels in spoken word processing has been widely investigated, and studies have found a phenomenon of a Consonantal bias (C-bias), indicating that consonants carry more weight than vowels. However, across languages, various patterns have been documented, including that of no preference or a reverse pattern of Vowel bias. A central question is how the manifestation of the C-bias is modulated by language-specific factors. This question can be addressed by cross-linguistic studies. Comparing native Hebrew and native English speakers, this study examines the relative importance of transitional probabilities between non-adjacent consonants as opposed to vowels during auditory statistical learning (SL) of an artificial language. Hebrew is interesting because its complex Semitic morphological structure has been found to play a central role in lexical access, allowing us to examine whether morphological properties can modulate the C-bias in early phases of speech perception, namely, word segmentation. As predicted, we found a significant interaction between language and consonant/vowel manipulation, with a higher performance in the consonantal condition than in the vowel condition for Hebrew speakers, namely, C-bias, and no consonant/vowel asymmetry among English speakers. We suggest that the observed interaction is morphologically anchored, indicating that phonological and morphological processes interact during early phases of auditory word perception.

9.
Phonetica ; 81(1): 1-41, 2024 Feb 26.
Article in English | MEDLINE | ID: mdl-38215190

ABSTRACT

We investigate how three adult groups - experienced L2 English listeners; experienced D2 (second dialect) listeners; and native L1/D1 listeners - categorise Australian English (AusE) lax front vowels /ɪ e æ/ in /hVt/, /hVl/ and /mVl/ environments in a forced-choice categorisation task of synthesised continua. In study 1, AusE listeners show predictable categorisations, with an effect of coarticulation raising the vowel in perception for nasal onset stimuli, and a following lateral lowering the vowel in perception. In study 2, Irish (D2) and Chinese listeners (L2) have different categorisations than AusE listeners, likely guided by their D1/L1. Coarticulation influences the D1/D2 groups in similar ways, but results in more difficulty and less agreement for the Chinese. We also investigate the role of extralinguistic factors. For the Chinese listeners, higher proficiency in English does not correlate with more Australian-like categorisation behaviour. However, having fewer Chinese in their social network results in more Australian-like categorisation for some stimuli. These findings lend partial support to the role of experience and exposure in L2/D2 contexts, whereby categorisation is likely still driven by native categories, with increased exposure leading to better mapping, but not to a restructuring of underlying phonetic categories.


Subject(s)
Speech Perception , Transients and Migrants , Adult , Humans , Australia , Language , Phonetics , China
10.
Int J Audiol ; 63(4): 260-268, 2024 Apr.
Article in English | MEDLINE | ID: mdl-36853200

ABSTRACT

OBJECTIVE: The study's objective was to identify consonant and vowel confusions in cochlear implant (CI) users, using a nonsense syllable repetition test. DESIGN: In this cross-sectional study, participants repeated recorded mono- and bisyllabic nonsense words and real-word monosyllables in an open-set design. STUDY SAMPLE: Twenty-eight Norwegian-speaking, well-performing adult CI users (13 unilateral and 15 bilateral), using implants from Cochlear, Med-El and Advanced Bionics, and a reference group of 20 listeners with normal hearing participated. RESULTS: For the CI users, consonants were confused more often than vowels (58% versus 71% correct). Voiced consonants were confused more often than unvoiced (54% versus 64% correct). Voiced stops were often repeated as unvoiced, whereas unvoiced stops were never repeated as voiced. The nasals were repeated correctly in one third of the cases and confused with other nasals in one third of the cases. The real-word monosyllable score was significantly higher than the nonsense syllable score (76% versus 63% correct). CONCLUSIONS: The study revealed a general devoicing bias for the stops and a high confusion rate of nasals with other nasals, which suggests that the low-frequency coding in CIs is insufficient. Furthermore, the nonsense syllable test exposed more perception errors than the real word test.


Subject(s)
Cochlear Implants , Speech Perception , Adult , Humans , Cross-Sectional Studies , Acoustic Stimulation , Phonetics
11.
Cortex ; 171: 287-307, 2024 Feb.
Article in English | MEDLINE | ID: mdl-38061210

ABSTRACT

The spectral formant structure and periodicity pitch are the major features that determine the identity of vowels and the characteristics of the speaker. However, very little is known about how the processing of these features in the auditory cortex changes during development. To address this question, we independently manipulated the periodicity and formant structure of vowels while measuring auditory cortex responses using magnetoencephalography (MEG) in children aged 7-12 years and adults. We analyzed the sustained negative shift of source current associated with these vowel properties, which was present in the auditory cortex in both age groups despite differences in the transient components of the auditory response. In adults, the sustained activation associated with formant structure was lateralized to the left hemisphere early in the auditory processing stream requiring neither attention nor semantic mapping. This lateralization was not yet established in children, in whom the right hemisphere contribution to formant processing was strong and decreased during or after puberty. In contrast to the formant structure, periodicity was associated with a greater response in the right hemisphere in both children and adults. These findings suggest that left-lateralization for the automatic processing of vowel formant structure emerges relatively late in ontogenesis and pose a serious challenge to current theories of hemispheric specialization for speech processing.


Subject(s)
Auditory Cortex , Speech Perception , Adult , Humans , Child , Auditory Cortex/physiology , Acoustic Stimulation , Auditory Perception/physiology , Magnetoencephalography , Speech/physiology , Speech Perception/physiology
12.
Int J Syst Evol Microbiol ; 73(12)2023 Dec.
Article in English | MEDLINE | ID: mdl-38054495

ABSTRACT

The International Code of Nomenclature of Prokaryotes (ICNP) provides guidance on the formation of names from compound words. This includes recommendations on the use of connecting vowels, which are meant to make names easier to pronounce. However, deployment of a connecting vowel when the preceding word element ends in the same vowel can make a name harder to spell or say, bringing us into conflict with the recommendations that we should avoid names that are disagreeable and difficult to pronounce. Given that there are many precedents where connecting vowels are not used in this context, particularly in names formed from the term 'alkali', I hereby propose an emendation to the ICNP to drop the connecting vowel when the preceding word element ends in the same vowel.


Subject(s)
Fatty Acids , Aprepitant , Phylogeny , Sequence Analysis, DNA , RNA, Ribosomal, 16S/genetics , DNA, Bacterial/genetics , Bacterial Typing Techniques , Base Composition , Fatty Acids/chemistry
13.
J Voice ; 2023 Nov 06.
Article in English | MEDLINE | ID: mdl-37940420

ABSTRACT

BACKGROUND: There have been numerous reports on the acoustic characteristics of singers' vowel articulation and phonation, and these studies cover many phonetic dimensions, such as fundamental frequency (F0), intensity, formant frequency, and voice quality. METHOD: Taking the three representative vowels (/a/, /i/, /u/) in Chinese National Singing and Bel Canto as the research object, the present study investigates the differences and associations in vowel articulation and phonation between Chinese National Singing and Bel Canto using acoustic measures, for example, F0, formant frequency, long-term average spectrum (LTAS). RESULTS: The relationship between F0 and formant indicates that F1 is proportional to F0, in which the female has a significant variation in vowel /a/. Compared with the male, the formant structure of the female singing voice differs significantly from that of the speech voice. Regarding the relationship between intensity and formant, LTAS shows that the Chinese National Singing tenor and Bel Canto baritone have the singer's formant cluster when singing vowels, while the two sopranos do not. CONCLUSIONS: The systematic changes of formant frequencies with voice source are observed. (i) F1 of the female vowel /a/ has undergone a significant tuning change in the register transition, reflecting the characteristics of singing genres. (ii) Female singers utilize the intrinsic pitch of vowels when adopting the register transition strategy. This finding can be assumed to facilitate understanding the theory of intrinsic vowel pitch and revise Sundberg's hypothesis that F1 rises with F0. A non-linear relationship exists between F1 and F0, which adds to the non-linear interaction of the formant and vocal source. (iii) The singer's formant is affected by voice classification, gender, and singing genres.

14.
J Voice ; 2023 Nov 23.
Article in English | MEDLINE | ID: mdl-38000960

ABSTRACT

This study introduces the weighted linear prediction adapted to high-pitched singing voices (WLP-HPSV) method for accurately estimating formant frequencies of vowels sung by lyric sopranos. The WLP-HPSV method employs a variant of the WLP analysis combined with the zero-frequency filtering (ZFF) technique to address specific challenges in formant estimation from singing signals. Evaluation of the WLP-HPSV method compared to the LPC method demonstrated its superior performance in accurately capturing the spectral characteristics of synthetic /u/ vowels and the /a/ and /u/ natural singing vowels. The QCP parameters used in the WLP-HPSV method varied with pitch, revealing insights into the interplay between the vocal tract and glottal characteristics during vowel production. The comparison between the LPC and WLP-HPSV methods highlighted the robustness of the WLP-HPSV method in accurately estimating formant frequencies across different pitches.

15.
J Voice ; 37(5): 648-662, 2023 Sep.
Article in English | MEDLINE | ID: mdl-37717981

ABSTRACT

OBJECTIVE: Speech signal processing has become an important technique to ensure that the voice interaction system communicates accurately with the user by improving the clarity or intelligibility of speech signals. However, most existing works only focus on whether to process the voice of average human but ignore the communication needs of individuals suffering from voice disorder, including voice-related professionals, older people, and smokers. To solve this demand, it is essential to design a non-invasive repair system that processes pathological voices. METHODS: In this paper, we propose a repair system for multiple polyp vowels, such as /a/, /i/ and /u/. We utilize a non-linear model based on amplitude-modulation (AM) and a frequency-modulation (FM) structure to extract the pitch and formant of pathological voice. To solve the fracture and instability of pitch, we provide a pitch extraction algorithm, which ensures that pitch's stability and avoids the errors of double pitch caused by the instability of low-frequency signal. Furthermore, we design a formant reconstruction mechanism, which can effectively determine the frequency and bandwidth to accomplish formant repair. RESULTS: Finally, spectrum observation and objective indicators show that the system has better performance in improving the intelligibility of pathological speech.


Subject(s)
Voice Disorders , Voice , Humans , Aged , Speech , Voice Disorders/diagnosis , Algorithms , Cognition
16.
Front Psychol ; 14: 1133859, 2023.
Article in English | MEDLINE | ID: mdl-37448717

ABSTRACT

A person's first language (L1) affects the way they acquire speech in a second language (L2). However, we know relatively little about the role different varieties of the L1 play in the acquisition of L2 speech. This study focuses on German (L1) learners of English (L2) and asks whether the degree to which German speakers distinguish between the two vowels /eː/ and /ɛː/ in their L1 has an impact on how well these individuals identify /æ/ and discriminate between the two English vowels /ɛ/ and /æ/. These two English vowels differ in both vowel quality and duration (/æ/ is longer than /ɛ/). We report on an identification and a discrimination experiment. In the first study, participants heard a sound file and were asked to indicate whether they heard "pen" or "pan" (or "pedal" or "paddle"). The stimuli differed from each other in terms of both vowel quality (11 steps on a spectral continuum from an extreme /æ/ to an extreme /ɛ/) and duration (short, middle, long). In the second study, participants had to signal whether two sound files they were exposed to differed from each other. We modeled the percentage of /æ/ ("pan," "paddle") selection (identification task only, binomial logistic regression), accuracy (discrimination task only, binomial logistic regression), and reaction time (identification and discrimination tasks, linear mixed effects models) by implementing the German Pillai score as a measure of vowel overlap in our analysis. Each participant has an individual Pillai score, which ranges from 0 (= merger of L1 German /eː/ and /ɛː/) to 1 (=maintenance of the contrast between L1 German /eː/ and /ɛː/) and had been established, prior to the perception experiments reported here, in a production study. Although the findings from the discrimination study remain inconclusive, the results from the identification test support the hypothesis that maintaining the vowel contrast in the L1 German leads to a more native-like identification of /æ/ in L2 English. We conclude that sociolinguistic variation in someone's L1 can affect L2 acquisition.

17.
Front Hum Neurosci ; 17: 1163578, 2023.
Article in English | MEDLINE | ID: mdl-37275343

ABSTRACT

Speech imagery recognition from electroencephalograms (EEGs) could potentially become a strong contender among non-invasive brain-computer interfaces (BCIs). In this report, first we extract language representations as the difference of line-spectra of phones by statistically analyzing many EEG signals from the Broca area. Then we extract vowels by using iterative search from hand-labeled short-syllable data. The iterative search process consists of principal component analysis (PCA) that visualizes linguistic representation of vowels through eigen-vectors φ(m), and subspace method (SM) that searches an optimum line-spectrum for redesigning φ(m). The extracted linguistic representation of Japanese vowels /i/ /e/ /a/ /o/ /u/ shows 2 distinguished spectral peaks (P1, P2) in the upper frequency range. The 5 vowels are aligned on the P1-P2 chart. A 5-vowel recognition experiment using a data set of 5 subjects and a convolutional neural network (CNN) classifier gave a mean accuracy rate of 72.6%.

18.
J Child Lang ; : 1-24, 2023 Jun 13.
Article in English | MEDLINE | ID: mdl-37309654

ABSTRACT

While adult studies show that consonants are more important than vowels in lexical processing tasks, the developmental trajectory of this consonant bias varies cross-linguistically. This study tested whether British English-learning 11-month-old infants' recognition of familiar word forms is more reliant on consonants than vowels, as found by Poltrock and Nazzi (2015) in French. After establishing that infants prefer listening to a list of familiar words over pseudowords (Experiment 1), Experiment 2 examined preference for consonant versus vowel mispronunciations of these words. Infants listened to both alterations equally. In Experiment 3, using a simplified version of the task with one familiar word only ('mummy'), infants' preference for its correct pronunciation over a consonant or a vowel change confirmed an equal sensitivity to both alterations. British English-learning infants' word form recognition appears to be equally impacted by consonant and vowel information, providing further evidence that initial lexical processes vary cross-linguistically.

19.
Phonetica ; 80(1-2): 1-42, 2023 02 23.
Article in English | MEDLINE | ID: mdl-37314963

ABSTRACT

Study 1 compared vowels in Child Directed Speech (CDS; child ages 25-46 months) to vowels in Adult Directed Speech (ADS) in natural conversation in the Australian Indigenous language Warlpiri, which has three vowels (/i/, /a/, /u). Study 2 compared the vowels of the child interlocutors from Study 1 to caregiver ADS and CDS. Study 1 indicates that Warlpiri CDS vowels are characterised by fronting, /a/-lowering, f o -raising, and increased duration, but not vowel space expansion. Vowels in CDS nouns, however, show increased between-contrast differentiation and reduced within-contrast variation, similar to what has been reported for other languages. We argue that this two-part CDS modification process serves a dual purpose: Vowel space shifting induces IDS/CDS that sounds more child-like, which may enhance child attention to speech, while increased between-contrast differentiation and reduced within-contrast variation in nouns may serve didactic purposes by providing high-quality information about lexical specifications. Study 2 indicates that Warlpiri CDS vowels are more like child vowels, providing indirect evidence that aspects of CDS may serve non-linguistic purposes simultaneously with other aspects serving linguistic-didactic purposes. The studies have novel implications for the way CDS vowel modifications are considered and highlight the necessity of naturalistic data collection, novel analyses, and typological diversity.


Subject(s)
Speech Perception , Speech , Adult , Humans , Phonetics , Australia , Language , Speech Acoustics
20.
Brain Sci ; 13(5)2023 May 16.
Article in English | MEDLINE | ID: mdl-37239282

ABSTRACT

The music and spoken language domains share acoustic properties such as fundamental frequency (f0, perceived as pitch), duration, resonance frequencies, and intensity. In speech, the acoustic properties form an essential part in differentiating between consonants, vowels, and lexical tones. This study investigated whether there is any advantage of musicality in the perception and production of Thai speech sounds. Two groups of English-speaking adults-one comprising formally trained musicians and the other non-musicians-were tested for their perception and production of Thai consonants, vowels, and tones. For both groups, the perception and production accuracy scores were higher for vowels than consonants and tones, and in production, there was also better accuracy for tones than consonants. Between the groups, musicians (defined as having more than five years of formal musical training) outperformed non-musicians (defined as having less than two years of formal musical training) in both the perception and production of all three sound types. Additional experiential factors that positively influenced the accuracy rates were the current hours of practice per week and those with some indication of an augmentation due to musical aptitude, but only in perception. These results suggest that music training, defined as formal training for more than five years, and musical training, expressed in hours of weekly practice, facilitate the perception and production of non-native speech sounds.

SELECTION OF CITATIONS
SEARCH DETAIL
...