Your browser doesn't support javascript.
loading
: 20 | 50 | 100
1 - 20 de 18.470
1.
J Acoust Soc Am ; 155(5): 3060-3070, 2024 May 01.
Article En | MEDLINE | ID: mdl-38717210

Speakers tailor their speech to different types of interlocutors. For example, speech directed to voice technology has different acoustic-phonetic characteristics than speech directed to a human. The present study investigates the perceptual consequences of human- and device-directed registers in English. We compare two groups of speakers: participants whose first language is English (L1) and bilingual L1 Mandarin-L2 English talkers. Participants produced short sentences in several conditions: an initial production and a repeat production after a human or device guise indicated either understanding or misunderstanding. In experiment 1, a separate group of L1 English listeners heard these sentences and transcribed the target words. In experiment 2, the same productions were transcribed by an automatic speech recognition (ASR) system. Results show that transcription accuracy was highest for L1 talkers for both human and ASR transcribers. Furthermore, there were no overall differences in transcription accuracy between human- and device-directed speech. Finally, while human listeners showed an intelligibility benefit for coda repair productions, the ASR transcriber did not benefit from these enhancements. Findings are discussed in terms of models of register adaptation, phonetic variation, and human-computer interaction.


Multilingualism , Speech Intelligibility , Speech Perception , Humans , Male , Female , Adult , Young Adult , Speech Acoustics , Phonetics , Speech Recognition Software
2.
J Acoust Soc Am ; 155(5): 3071-3089, 2024 May 01.
Article En | MEDLINE | ID: mdl-38717213

This study investigated how 40 Chinese learners of English as a foreign language (EFL learners) differed from 40 native English speakers in the production of four English tense-lax contrasts, /i-ɪ/, /u-ʊ/, /ɑ-ʌ/, and /æ-ε/, by examining the acoustic measurements of duration, the first three formant frequencies, and the slope of the first formant movement (F1 slope). The dynamic formant trajectory was modeled using discrete cosine transform coefficients to demonstrate the time-varying properties of formant trajectories. A discriminant analysis was employed to illustrate the extent to which Chinese EFL learners relied on different acoustic parameters. This study found that: (1) Chinese EFL learners overemphasized durational differences and weakened spectral differences for the /i-ɪ/, /u-ʊ/, and /ɑ-ʌ/ pairs, although they maintained sufficient spectral differences for /æ-ε/. In contrast, native English speakers predominantly used spectral differences across all four pairs; (2) in non-low tense-lax contrasts, unlike native English speakers, Chinese EFL learners failed to exhibit different F1 slope values, indicating a non-nativelike tongue-root placement during the articulatory process. The findings underscore the contribution of dynamic spectral patterns to the differentiation between English tense and lax vowels, and reveal the influence of precise articulatory gestures on the realization of the tense-lax contrast.


Multilingualism , Phonetics , Speech Acoustics , Humans , Male , Female , Young Adult , Speech Production Measurement , Adult , Language , Acoustics , Learning , Voice Quality , Sound Spectrography , East Asian People
3.
JASA Express Lett ; 4(5)2024 May 01.
Article En | MEDLINE | ID: mdl-38717469

The perceptual boundary between short and long categories depends on speech rate. We investigated the influence of speech rate on perceptual boundaries for short and long vowel and consonant contrasts by Spanish-English bilingual listeners and English monolinguals. Listeners tended to adapt their perceptual boundaries to speech rates, but the strategy differed between groups, especially for consonants. Understanding the factors that influence auditory processing in this population is essential for developing appropriate assessments of auditory comprehension. These findings have implications for the clinical care of older populations whose ability to rely on spectral and/or temporal information in the auditory signal may decline.


Multilingualism , Speech Perception , Humans , Speech Perception/physiology , Female , Male , Adult , Phonetics , Young Adult
4.
Dyslexia ; 30(3): e1774, 2024 Aug.
Article En | MEDLINE | ID: mdl-38807032

Working memory (WM) has been consistently linked to reading. However, the mechanism(s) linking WM to reading remain unclear. WM may indirectly exert an effect onto reading through mediators such as phonemic awareness (PA) and/or rapid automatized naming (RAN). In a sample of children with reading difficulty (n = 117), separate mediation analyses tested direct and indirect (through PA and RAN) effects of WM on untimed word decoding and recognition (i.e., basic reading skills) and timed word decoding and recognition (i.e., reading fluency). WM exerted a direct effect on basic reading skills and reading fluency. For basic reading skills, there was a significant indirect effect of WM on reading through the mediation of PA (but not through RAN). By contrast, for reading fluency, there was a significant indirect effect of WM on reading through the mediation of RAN (but not through PA). Findings reinforce the importance of WM, PA, and RAN for broad reading skills, while offering a mechanistic explanation for why poor PA and/or RAN may differentially lead to reading difficulty.


Awareness , Dyslexia , Memory, Short-Term , Phonetics , Reading , Humans , Memory, Short-Term/physiology , Child , Female , Male , Dyslexia/physiopathology , Awareness/physiology
5.
J Acoust Soc Am ; 155(5): 3521-3536, 2024 May 01.
Article En | MEDLINE | ID: mdl-38809098

This electromagnetic articulography study explores the kinematic profile of Intonational Phrase boundaries in Seoul Korean. Recent findings suggest that the scope of phrase-final lengthening is conditioned by word- and/or phrase-level prominence. However, evidence comes mainly from head-prominence languages, which conflate positions of word prosody with positions of phrasal prominence. Here, we examine phrase-final lengthening in Seoul Korean, an edge-prominence language with no word prosody, with respect to focus location as an index of phrase-level prominence and Accentual Phrase (AP) length as an index of word demarcation. Results show that phrase-final lengthening extends over the phrase-final syllable. The effect is greater the further away that focus occurs. It also interacts with the domains of AP and prosodic word: lengthening is greater in smaller APs, whereas shortening is observed in the initial gesture of the phrase-final word. Additional analyses of kinematic displacement and peak velocity revealed that Korean phrase-final gestures bear the kinematic profile of IP boundaries concurrently to what is typically considered prominence marking. Based on these results, a gestural coordination account is proposed, in which boundary-related events interact systematically with phrase-level prominence as well as lower prosodic levels, and how this proposal relates to the findings in head-prominence languages is discussed.


Phonetics , Speech Acoustics , Humans , Male , Female , Young Adult , Biomechanical Phenomena , Adult , Language , Gestures , Speech Production Measurement , Republic of Korea , Voice Quality , Time Factors
6.
Lang Speech ; 67(2): 279-300, 2024 Jun.
Article En | MEDLINE | ID: mdl-38756046

The paper introduces the Special Issue on Language Contact and Speaker Accommodation, which originates from the conference Phonetics and Phonology in Europe (PaPE) held at the University of Lecce, Italy, in 2019. It discusses the topics of language contact and speaker accommodation, summarizing the contributions included in the Special Issue, and arguing explicitly in favour of a unitary view of how both temporary and stable changes happen in (part of) the linguistic systems. Accommodation is seen as the same gradual and non-homogeneous process at play in different contact settings. In the introductory sections, a discussion is offered on various situations in which linguistic systems are in contact and on the main factors that may be at play; the following sections offer an overview of the papers included in the Special Issue, which focus on accommodation in L2 and heritage speakers as well as on the time dimension of dialect or language societal contact. Finally, accommodation is discussed as the same process that is at work in any interaction, that may modify temporarily or long-term the system of L2 learners and bilinguals (e.g., immigrants), that usually affects in the long-term the heritage speakers' system, and that only in the long term can lead to language changes involving entire communities.


Multilingualism , Humans , Language , Phonetics , Speech
7.
JASA Express Lett ; 4(5)2024 May 01.
Article En | MEDLINE | ID: mdl-38804812

Adding to limited research on clear speech in tone languages, productions of Mandarin lexical tones were examined in pentasyllabic sentences. Fourteen participants read sentences imagining a hard-of-hearing addressee or a friend in a casual social setting. Tones produced in clear speech had longer duration, higher intensity, and larger F0 values. This style effect was rarely modulated by tone, preceding tonal context, or syllable position, consistent with an overall signal enhancement strategy. Possible evidence for tone enhancement was observed only in one set of analysis for F0 minimum and F0 range, contrasting tones with low targets and tones with high targets.


Language , Humans , Female , Male , Speech Acoustics , Adult , Young Adult , Speech , Speech Perception/physiology , Phonetics
8.
PLoS One ; 19(5): e0304040, 2024.
Article En | MEDLINE | ID: mdl-38814896

This study investigates head nods in natural dyadic German Sign Language (DGS) interaction, with the aim of finding whether head nods serving different functions vary in their phonetic characteristics. Earlier research on spoken and sign language interaction has revealed that head nods vary in the form of the movement. However, most claims about the phonetic properties of head nods have been based on manual annotation without reference to naturalistic text types and the head nods produced by the addressee have been largely ignored. There is a lack of detailed information about the phonetic properties of the addressee's head nods and their interaction with manual cues in DGS as well as in other sign languages, and the existence of a form-function relationship of head nods remains uncertain. We hypothesize that head nods functioning in the context of affirmation differ from those signaling feedback in their form and the co-occurrence with manual items. To test the hypothesis, we apply OpenPose, a computer vision toolkit, to extract head nod measurements from video recordings and examine head nods in terms of their duration, amplitude and velocity. We describe the basic phonetic properties of head nods in DGS and their interaction with manual items in naturalistic corpus data. Our results show that phonetic properties of affirmative nods differ from those of feedback nods. Feedback nods appear to be on average slower in production and smaller in amplitude than affirmation nods, and they are commonly produced without a co-occurring manual element. We attribute the variations in phonetic properties to the distinct roles these cues fulfill in turn-taking system. This research underlines the importance of non-manual cues in shaping the turn-taking system of sign languages, establishing the links between such research fields as sign language linguistics, conversational analysis, quantitative linguistics and computer vision.


Phonetics , Sign Language , Humans , Germany , Male , Head/physiology , Female , Language , Head Movements/physiology
9.
PLoS One ; 19(5): e0304150, 2024.
Article En | MEDLINE | ID: mdl-38805447

When comprehending speech, listeners can use information encoded in visual cues from a face to enhance auditory speech comprehension. For example, prior work has shown that the mouth movements reflect articulatory features of speech segments and durational information, while pitch and speech amplitude are primarily cued by eyebrow and head movements. Little is known about how the visual perception of segmental and prosodic speech information is influenced by linguistic experience. Using eye-tracking, we studied how perceivers' visual scanning of different regions on a talking face predicts accuracy in a task targeting both segmental versus prosodic information, and also asked how this was influenced by language familiarity. Twenty-four native English perceivers heard two audio sentences in either English or Mandarin (an unfamiliar, non-native language), which sometimes differed in segmental or prosodic information (or both). Perceivers then saw a silent video of a talking face, and judged whether that video matched either the first or second audio sentence (or whether both sentences were the same). First, increased looking to the mouth predicted correct responses only for non-native language trials. Second, the start of a successful search for speech information in the mouth area was significantly delayed in non-native versus native trials, but just when there were only prosodic differences in the auditory sentences, and not when there were segmental differences. Third, (in correct trials) the saccade amplitude in native language trials was significantly greater than in non-native trials, indicating more intensely focused fixations in the latter. Taken together, these results suggest that mouth-looking was generally more evident when processing a non-native versus native language in all analyses, but fascinatingly, when measuring perceivers' latency to fixate the mouth, this language effect was largest in trials where only prosodic information was useful for the task.


Language , Phonetics , Speech Perception , Humans , Female , Male , Adult , Speech Perception/physiology , Young Adult , Face/physiology , Visual Perception/physiology , Eye Movements/physiology , Speech/physiology , Eye-Tracking Technology
10.
J Acoust Soc Am ; 155(5): 3206-3212, 2024 May 01.
Article En | MEDLINE | ID: mdl-38738937

Modern humans and chimpanzees share a common ancestor on the phylogenetic tree, yet chimpanzees do not spontaneously produce speech or speech sounds. The lab exercise presented in this paper was developed for undergraduate students in a course entitled "What's Special About Human Speech?" The exercise is based on acoustic analyses of the words "cup" and "papa" as spoken by Viki, a home-raised, speech-trained chimpanzee, as well as the words spoken by a human. The analyses allow students to relate differences in articulation and vocal abilities between Viki and humans to the known anatomical differences in their vocal systems. Anatomical and articulation differences between humans and Viki include (1) potential tongue movements, (2) presence or absence of laryngeal air sacs, (3) presence or absence of vocal membranes, and (4) exhalation vs inhalation during production.


Pan troglodytes , Speech Acoustics , Speech , Humans , Animals , Pan troglodytes/physiology , Speech/physiology , Tongue/physiology , Tongue/anatomy & histology , Vocalization, Animal/physiology , Species Specificity , Speech Production Measurement , Larynx/physiology , Larynx/anatomy & histology , Phonetics
11.
J Psycholinguist Res ; 53(4): 49, 2024 May 24.
Article En | MEDLINE | ID: mdl-38782761

Previous studies on L2 (i.e., second language) Chinese compound processing have focused on the relative efficiency of two routes: holistic processing versus combinatorial processing. However, it is still unclear whether Chinese compounds are processed with multilevel representations among L2 learners due to the hierarchical structure of the characters. Therefore, taking a multivariate approach, the present study evaluated the relative influence and importance of different grain sizes of lexical information in an L2 Chinese two-character compound decision task. Results of supervised component generalized linear regression models with random forests analysis revealed that the orthographic, phonological and semantic information all contributed to L2 compound processing, but the L2 learners used more orthographic processing strategies and fewer phonological processing strategies compared to the native speakers. Specifically, the orthographic information was activated at the whole-word, the character and the radical levels in orthographic processing, and the phonological information at the whole-word, the syllable, and the phoneme levels all exerted contributions in phonological processing. Furthermore, the semantic information of the whole words and the constituents was accessed in semantic processing. These findings together suggest that the L2 learners are able to use cues at all levels simultaneously to process Chinese compound words, supporting a multi-route model with a hierarchical morphological structure in such processing.


Multilingualism , Psycholinguistics , Semantics , Adult , Female , Humans , Male , Young Adult , China , Language , Phonetics , Reading
12.
Codas ; 36(3): e20220336, 2024.
Article En | MEDLINE | ID: mdl-38747755

PURPOSE: To review studies that have intervention in reading with impacts on phonological awareness in children with autism spectrum disorder. RESEARCH STRATEGIES: Searches took place until February 2021 in Cochrane, Embase, ERIC (Education Resources Information Center), LILACS (Latin American and Caribbean Health Sciences Literature), PubMed/Medline, Scopus, Web of Science and gray literature databases. SELECTION CRITERIA: The review included experimental studies with preschoolers and schoolchildren with ASD. Two independent reviewers selected the studies and, in case of disagreement, a third reviewer was consulted. DATA ANALYSIS: Joanna Briggs Institute checklists were used for risk of bias. A random effects meta-analysis was performed and the certainty of the evidence was assessed using the GRADE tool. RESULTS: Eight studies with some impact on phonological awareness were reviewed. The risk of bias was low and moderate. The certainty of the evidence was low for randomized trials and very low for non-randomised trials. Comparison of pre- and post-therapy on the Preschool Literacy Test (TOPEL) showed that children with ASD improved phonological awareness, with a mean difference between baseline and post-therapy of 6.21 (95% CI = 3.75-8.67; I2 = 0%). CONCLUSION: Shared reading and software activities with words and phrases can alter phonological awareness. These results support further research with larger samples and a detailed description of the intervention to observe its effectiveness in phonological awareness.


Autism Spectrum Disorder , Reading , Humans , Child , Child, Preschool , Phonetics
13.
J Psycholinguist Res ; 53(3): 45, 2024 May 13.
Article En | MEDLINE | ID: mdl-38739304

English is widely regarded as a global language, and it has become increasingly important for global communication. As a result, the demand for English language education has been on the rise. In China, a significant number of individuals are engaged in learning the English language. However, many English learners in China encounter challenges when it comes to developing their speaking skills. This study aims to investigate the factors influencing the speaking skills of English learners in China. Employing a mixed-methods approach, data were collected through a questionnaire from 455 college students from three different courses (arts, science & business, and commerce) in China. The study findings identified several factors impacting the speaking skills of English learners in China, including limited opportunities for speaking practice, fear of making mistakes, limited exposure to English-speaking environments, inadequate teacher training, and the influence of the Chinese language on English pronunciation. Additionally, the study highlighted that learners who have greater exposure to English-speaking environments and more opportunities for speaking practice tend to demonstrate better speaking skills. The novelty of this study lies in its valuable insights into the factors influencing the speaking skills of English learners in China. Based on the findings, it is recommended that English teachers receive enhanced training to effectively teach speaking skills, and learners should be provided with increased opportunities for speaking practice, such as participating in group discussions or engaging in speaking activities.


Learning , Humans , China , Female , Male , Learning/physiology , Young Adult , Multilingualism , Speech , Language , Students/psychology , Adult , Surveys and Questionnaires , Phonetics , East Asian People
14.
Sci Rep ; 14(1): 10249, 2024 05 04.
Article En | MEDLINE | ID: mdl-38704429

Phonological awareness (PA) is at the foundation of reading development: PA is introduced before formal reading instruction, predicts reading development, is a target for early intervention, and is a core mechanism in dyslexia. Conventional approaches to assessing PA are time-consuming and resource intensive: assessments are individually administered and scoring verbal responses is challenging and subjective. Therefore, we introduce a rapid, automated, online measure of PA-The Rapid Online Assessment of Reading-Phonological Awareness-that can be implemented at scale without a test administrator. We explored whether this gamified, online task is an accurate and reliable measure of PA and predicts reading development. We found high correlations with standardized measures of PA (CTOPP-2, r = .80) for children from Pre-K through fourth grade and exceptional reliability (α = .96). Validation in 50 first and second grade classrooms showed reliable implementation in a public school setting with predictive value of future reading development.


Dyslexia , Phonetics , Reading , Humans , Child , Female , Male , Dyslexia/diagnosis , Dyslexia/physiopathology , Reproducibility of Results , Awareness , Child, Preschool
15.
J Commun Disord ; 109: 106428, 2024.
Article En | MEDLINE | ID: mdl-38744198

PURPOSE: This study examines whether there are differences in the speech of speakers with dysarthria, speakers with apraxia and healthy speakers in spectral acoustic measures during production of the central-peninsular Spanish alveolar sibilant fricative /s/. METHOD: To this end, production of the sibilant was analyzed in 20 subjects with dysarthria, 8 with apraxia of speech and 28 healthy speakers. Participants produced 12 sV(C) words. The variables compared across groups were the fricative's spectral amplitude difference (AmpD) and spectral moments in the temporal midpoint of fricative execution. RESULTS: The results indicate that individuals with dysarthria can be distinguished from healthy speakers in terms of the spectral characteristics AmpD, standard deviation (SD), center of gravity (CoG) and skewness, the last two in context with unrounded vowel, while no differences in kurtosis were detected. Participants with AoS group differ significantly from healthy speaker group in AmpD, SD and CoG and Kurtosis, the first one followed unrounded vowel and the latter two followed by rounded vowels. In addition, speakers with apraxia of speech group returned significant differences with respect to speakers with dysarthria group in AmpD, CoG and skewness. CONCLUSIONS: The differences found between the groups in the measures studied as a function of the type of vowel context could provide insights into the distinctive manifestations of motor speech disorders, contributing to the differential diagnosis between apraxia and dysarthria in motor control processes.


Apraxias , Dysarthria , Speech Acoustics , Humans , Dysarthria/physiopathology , Dysarthria/etiology , Apraxias/physiopathology , Male , Female , Middle Aged , Adult , Aged , Phonetics , Speech Production Measurement
16.
PLoS One ; 19(5): e0299140, 2024.
Article En | MEDLINE | ID: mdl-38809807

Non-random exploration of infant speech-like vocalizations (e.g., squeals, growls, and vowel-like sounds or "vocants") is pivotal in speech development. This type of vocal exploration, often noticed when infants produce particular vocal types in clusters, serves two crucial purposes: it establishes a foundation for speech because speech requires formation of new vocal categories, and it serves as a basis for vocal signaling of wellness and interaction with caregivers. Despite the significance of clustering, existing research has largely relied on subjective descriptions and anecdotal observations regarding early vocal category formation. In this study, we aim to address this gap by presenting the first large-scale empirical evidence of vocal category exploration and clustering throughout the first year of life. We observed infant vocalizations longitudinally using all-day home recordings from 130 typically developing infants across the entire first year of life. To identify clustering patterns, we conducted Fisher's exact tests to compare the occurrence of squeals versus vocants, as well as growls versus vocants. We found that across the first year, infants demonstrated clear clustering patterns of squeals and growls, indicating that these categories were not randomly produced, but rather, it seemed, infants actively engaged in practice of these specific categories. The findings lend support to the concept of infants as manifesting active vocal exploration and category formation, a key foundation for vocal language.


Speech , Humans , Infant , Male , Female , Speech/physiology , Language Development , Voice/physiology , Longitudinal Studies , Phonetics
18.
J Acoust Soc Am ; 155(4): 2285-2301, 2024 Apr 01.
Article En | MEDLINE | ID: mdl-38557735

Fronting of the vowels /u, ʊ, o/ is observed throughout most North American English varieties, but has been analyzed mainly in terms of acoustics rather than articulation. Because an increase in F2, the acoustic correlate of vowel fronting, can be the result of any gesture that shortens the front cavity of the vocal tract, acoustic data alone do not reveal the combination of tongue fronting and/or lip unrounding that speakers use to produce fronted vowels. It is furthermore unresolved to what extent the articulation of fronted back vowels varies according to consonantal context and how the tongue and lips contribute to the F2 trajectory throughout the vowel. This paper presents articulatory and acoustic data on fronted back vowels from two varieties of American English: coastal Southern California and South Carolina. Through analysis of dynamic acoustic, ultrasound, and lip video data, it is shown that speakers of both varieties produce fronted /u, ʊ, o/ with rounded lips, and that high F2 observed for these vowels is associated with a front-central tongue position rather than unrounded lips. Examination of time-varying formant trajectories and articulatory configurations shows that the degree of vowel-internal F2 change is predominantly determined by coarticulatory influence of the coda.


Phonetics , Speech Acoustics , United States , Acoustics , Language , South Carolina
19.
J Acoust Soc Am ; 155(4): 2698-2706, 2024 Apr 01.
Article En | MEDLINE | ID: mdl-38639561

The notion of the "perceptual center" or the "P-center" has been put forward to account for the repeated finding that acoustic and perceived syllable onsets do not necessarily coincide, at least in the perception of simple monosyllables or disyllables. The magnitude of the discrepancy between acoustics and perception-the location of the P-center in the speech signal- has proven difficult to estimate, though acoustic models of the effect do exist. The present study asks if the P-center effect can be documented in natural connected speech of English and Japanese and examines if an acoustic model that defines the P-center as the moment of the fastest energy change in a syllabic amplitude envelope adequately reflects the P-center in the two languages. A sensorimotor synchronization paradigm was deployed to address the research questions. The results provide evidence for the existence of the P-center effect in speech of both languages while the acoustic P-center model is found to be less applicable to Japanese. Sensorimotor synchronization patterns further suggest that the P-center may reflect perceptual anticipation of a vowel onset.


Speech Acoustics , Speech Perception , Humans , Phonetics , Speech , Language
20.
J Acoust Soc Am ; 155(4): 2687-2697, 2024 Apr 01.
Article En | MEDLINE | ID: mdl-38639927

One speech sound can be associated with multiple meanings through iconicity, indexicality, and/or systematicity. It was not until recently that this "pluripotentiality" of sound symbolism attracted serious attention, and it remains uninvestigated how pluripotentiality may arise. In the current study, Japanese, Korean, Mandarin, and English speakers rated unfamiliar jewel names on three semantic scales: size, brightness, and hardness. The results showed language-specific and cross-linguistically shared pluripotential sound symbolism. Japanese speakers associated voiced stops with large and dark jewels, whereas Mandarin speakers associated [i] with small and bright jewels. Japanese, Mandarin, and English speakers also associated lip rounding with darkness and softness. These sound-symbolic meanings are unlikely to be obtained through metaphorical or metonymical extension, nor are they reported to colexify. Notably, in a purely semantic network without the mediation of lip rounding, softness can instead be associated with brightness, as illustrated by synesthetic metaphors such as yawaraka-na hizashi /jawaɾakanaçizaɕi/ "a gentle (lit. soft) sunshine" in Japanese. These findings suggest that the semantic networks of sound symbolism may not coincide with those of metaphor or metonymy. The current study summarizes the findings in the form of (phono)semantic maps to facilitate cross-linguistic comparisons of pluripotential sound symbolism.


Language , Semantic Web , Symbolism , Semantics , Phonetics
...