Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 22
Filtrar
Más filtros










Base de datos
Intervalo de año de publicación
1.
Sci Rep ; 14(1): 4852, 2024 02 28.
Artículo en Inglés | MEDLINE | ID: mdl-38418850

RESUMEN

Assessing infant carrying and holding (C/H), or physical infant-caregiver interaction, is important for a wide range of contexts in development research. An automated detection and quantification of infant C/H is particularly needed in long term at-home studies where development of infants' neurobehavior is measured using wearable devices. Here, we first developed a phenomenological categorization for physical infant-caregiver interactions to support five different definitions of C/H behaviors. Then, we trained and assessed deep learning-based classifiers for their automatic detection from multi-sensor wearable recordings that were originally used for mobile assessment of infants' motor development. Our results show that an automated C/H detection is feasible at few-second temporal accuracy. With the best C/H definition, the automated detector shows 96% accuracy and 0.56 kappa, which is slightly less than the video-based inter-rater agreement between trained human experts (98% accuracy, 0.77 kappa). The classifier performance varies with C/H definition reflecting the extent to which infants' movements are present in each C/H variant. A systematic benchmarking experiment shows that the widely used actigraphy-based method ignores the normally occurring C/H behaviors. Finally, we show proof-of-concept for the utility of the novel classifier in studying C/H behavior across infant development. Particularly, we show that matching the C/H detections to individuals' gross motor ability discloses novel insights to infant-parent interaction.


Asunto(s)
Movimiento , Dispositivos Electrónicos Vestibles , Lactante , Niño , Humanos , Desarrollo Infantil , Actigrafía , Padres
2.
Artículo en Inglés | MEDLINE | ID: mdl-38083169

RESUMEN

The recently-developed infant wearable MAIJU provides a means to automatically evaluate infants' motor performance in an objective and scalable manner in out-of-hospital settings. This information could be used for developmental research and to support clinical decision-making, such as detection of developmental problems and guiding of their therapeutic interventions. MAIJU-based analyses rely fully on the classification of infant's posture and movement; it is hence essential to study ways to increase the accuracy of such classifications, aiming to increase the reliability and robustness of the automated analysis. Here, we investigated how self-supervised pre-training improves performance of the classifiers used for analyzing MAIJU recordings, and we studied whether performance of the classifier models is affected by context-selective quality-screening of pre-training data to exclude periods of little infant movement or with missing sensors. Our experiments show that i) pre-training the classifier with unlabeled data leads to a robust accuracy increase of subsequent classification models, and ii) selecting context-relevant pre-training data leads to substantial further improvements in the classifier performance.Clinical relevance- This study showcases that self-supervised learning can be used to increase the accuracy of out-of-hospital evaluation of infants' motor abilities via smart wearables.


Asunto(s)
Movimiento , Dispositivos Electrónicos Vestibles , Lactante , Humanos , Reproducibilidad de los Resultados , Postura
3.
Int J Speech Lang Pathol ; : 1-11, 2023 Oct 06.
Artículo en Inglés | MEDLINE | ID: mdl-37800979

RESUMEN

PURPOSE: The purpose of this study was to analyse the relationship between automatic vowel articulation index (aVAI) and direct magnitude estimation (DME) among speakers with Parkinson's disease (PD) and healthy controls. We further analysed the potential of aVAI to serve as an objective measure of speech impairment in the clinical setting. METHOD: Speech samples from native Finnish speakers were utilised. Expert raters utilised DME to scale the intelligibility of speech samples. aVAI scores for PD speakers and healthy control speakers were analysed in relationship to DME speech intelligibility ratings and, among PD speakers, disease stage utilising nonparametric statistical analysis. RESULT: Mean DME intelligibility ratings were lower among PD speakers compared to healthy controls. Mean aVAI scores were nearly the same between speaker groups. DME intelligibility ratings and aVAI were strongly correlated within the PD speaker group. aVAI and DME intelligibility ratings were moderately correlated with disease stage as measured by the Hoehn and Yahr scale. CONCLUSION: aVAI was observed to be a promising tool for analysing vowel articulation in PD speakers. Further research is warranted on the application of aVAI as an objective measure of severity of speech impairment in the clinical setting, with varying patient populations and speech samples.

4.
Cogn Sci ; 47(7): e13307, 2023 07.
Artículo en Inglés | MEDLINE | ID: mdl-37395673

RESUMEN

Computational models of child language development can help us understand the cognitive underpinnings of the language learning process, which occurs along several linguistic levels at once (e.g., prosodic and phonological). However, in light of the replication crisis, modelers face the challenge of selecting representative and consolidated infant data. Thus, it is desirable to have evaluation methodologies that could account for robust empirical reference data, across multiple infant capabilities. Moreover, there is a need for practices that can compare developmental trajectories of infants to those of models as a function of language experience and development. The present study aims to take concrete steps to address these needs by introducing the concept of comparing models with large-scale cumulative empirical data from infants, as quantified by meta-analyses conducted across a large number of individual behavioral studies. We formalize the connection between measurable model and human behavior, and then present a conceptual framework for meta-analytic evaluation of computational models. We exemplify the meta-analytic model evaluation approach with two modeling experiments on infant-directed speech preference and native/non-native vowel discrimination.


Asunto(s)
Percepción del Habla , Humanos , Lactante , Simulación por Computador , Lenguaje , Desarrollo del Lenguaje , Habla
5.
EBioMedicine ; 92: 104591, 2023 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-37137181

RESUMEN

BACKGROUND: Early neurodevelopmental care and research are in urgent need of practical methods for quantitative assessment of early motor development. Here, performance of a wearable system in early motor assessment was validated and compared to developmental tracking of physical growth charts. METHODS: Altogether 1358 h of spontaneous movement during 226 recording sessions in 116 infants (age 4-19 months) were analysed using a multisensor wearable system. A deep learning-based automatic pipeline quantified categories of infants' postures and movements at a time scale of seconds. Results from an archived cohort (dataset 1, N = 55 infants) recorded under partial supervision were compared to a validation cohort (dataset 2, N = 61) recorded at infants' homes by the parents. Aggregated recording-level measures including developmental age prediction (DAP) were used for comparison between cohorts. The motor growth was also compared with respective DAP estimates based on physical growth data (length, weight, and head circumference) obtained from a large cohort (N = 17,838 infants; age 4-18 months). FINDINGS: Age-specific distributions of posture and movement categories were highly similar between infant cohorts. The DAP scores correlated tightly with age, explaining 97-99% (94-99% CI 95) of the variance at the group average level, and 80-82% (72-88%) of the variance in the individual recordings. Both the average motor and the physical growth measures showed a very strong fit to their respective developmental models (R2 = 0.99). However, single measurements showed more modality-dependent variation that was lowest for motor (σ = 1.4 [1.3-1.5 CI 95] months), length (σ = 1.5 months), and combined physical (σ = 1.5 months) measurements, and it was clearly higher for the weight (σ = 1.9 months) and head circumference (σ = 1.9 months) measurements. Longitudinal tracking showed clear individual trajectories, and its accuracy was comparable between motor and physical measures with longer measurement intervals. INTERPRETATION: A quantified, transparent and explainable assessment of infants' motor performance is possible with a fully automated analysis pipeline, and the results replicate across independent cohorts from out-of-hospital recordings. A holistic assessment of motor development provides an accuracy that is comparable with the conventional physical growth measures. A quantitative measure of infants' motor development may directly support individual diagnostics and care, as well as facilitate clinical research as an outcome measure in early intervention trials. FUNDING: This work was supported by the Finnish Academy (314602, 335788, 335872, 332017, 343498), Finnish Pediatric Foundation (Lastentautiensäätiö), Aivosäätiö, Sigrid Jusélius Foundation, and HUS Children's Hospital/HUS diagnostic center research funds.


Asunto(s)
Desarrollo Infantil , Dispositivos Electrónicos Vestibles , Lactante , Humanos , Niño , Gráficos de Crecimiento , Postura
6.
Sensors (Basel) ; 23(7)2023 Apr 06.
Artículo en Inglés | MEDLINE | ID: mdl-37050833

RESUMEN

Infant motility assessment using intelligent wearables is a promising new approach for assessment of infant neurophysiological development, and where efficient signal analysis plays a central role. This study investigates the use of different end-to-end neural network architectures for processing infant motility data from wearable sensors. We focus on the performance and computational burden of alternative sensor encoder and time series modeling modules and their combinations. In addition, we explore the benefits of data augmentation methods in ideal and nonideal recording conditions. The experiments are conducted using a dataset of multisensor movement recordings from 7-month-old infants, as captured by a recently proposed smart jumpsuit for infant motility assessment. Our results indicate that the choice of the encoder module has a major impact on classifier performance. For sensor encoders, the best performance was obtained with parallel two-dimensional convolutions for intrasensor channel fusion with shared weights for all sensors. The results also indicate that a relatively compact feature representation is obtainable for within-sensor feature extraction without a drastic loss to classifier performance. Comparison of time series models revealed that feedforward dilated convolutions with residual and skip connections outperformed all recurrent neural network (RNN)-based models in performance, training time, and training stability. The experiments also indicate that data augmentation improves model robustness in simulated packet loss or sensor dropout scenarios. In particular, signal- and sensor-dropout-based augmentation strategies provided considerable boosts to performance without negatively affecting the baseline performance. Overall, the results provide tangible suggestions on how to optimize end-to-end neural network training for multichannel movement sensor data.


Asunto(s)
Redes Neurales de la Computación , Dispositivos Electrónicos Vestibles , Humanos , Lactante , Movimiento/fisiología
7.
Commun Med (Lond) ; 2: 69, 2022.
Artículo en Inglés | MEDLINE | ID: mdl-35721830

RESUMEN

Background: Early neurodevelopmental care needs better, effective and objective solutions for assessing infants' motor abilities. Novel wearable technology opens possibilities for characterizing spontaneous movement behavior. This work seeks to construct and validate a generalizable, scalable, and effective method to measure infants' spontaneous motor abilities across all motor milestones from lying supine to fluent walking. Methods: A multi-sensor infant wearable was constructed, and 59 infants (age 5-19 months) were recorded during their spontaneous play. A novel gross motor description scheme was used for human visual classification of postures and movements at a second-level time resolution. A deep learning -based classifier was then trained to mimic human annotations, and aggregated recording-level outputs were used to provide posture- and movement-specific developmental trajectories, which enabled more holistic assessments of motor maturity. Results: Recordings were technically successful in all infants, and the algorithmic analysis showed human-equivalent-level accuracy in quantifying the observed postures and movements. The aggregated recordings were used to train an algorithm for predicting a novel neurodevelopmental measure, Baba Infant Motor Score (BIMS). This index estimates maturity of infants' motor abilities, and it correlates very strongly (Pearson's r = 0.89, p < 1e-20) to the chronological age of the infant. Conclusions: The results show that out-of-hospital assessment of infants' motor ability is possible using a multi-sensor wearable. The algorithmic analysis provides metrics of motility that are transparent, objective, intuitively interpretable, and they link strongly to infants' age. Such a solution could be automated and scaled to a global extent, holding promise for functional benchmarking in individualized patient care or early intervention trials.

8.
Behav Res Methods ; 53(2): 467-486, 2021 04.
Artículo en Inglés | MEDLINE | ID: mdl-32728916

RESUMEN

In the previous decade, dozens of studies involving thousands of children across several research disciplines have made use of a combined daylong audio-recorder and automated algorithmic analysis called the LENAⓇ system, which aims to assess children's language environment. While the system's prevalence in the language acquisition domain is steadily growing, there are only scattered validation efforts on only some of its key characteristics. Here, we assess the LENAⓇ system's accuracy across all of its key measures: speaker classification, Child Vocalization Counts (CVC), Conversational Turn Counts (CTC), and Adult Word Counts (AWC). Our assessment is based on manual annotation of clips that have been randomly or periodically sampled out of daylong recordings, collected from (a) populations similar to the system's original training data (North American English-learning children aged 3-36 months), (b) children learning another dialect of English (UK), and (c) slightly older children growing up in a different linguistic and socio-cultural setting (Tsimane' learners in rural Bolivia). We find reasonably high accuracy in some measures (AWC, CVC), with more problematic levels of performance in others (CTC, precision of male adults and other children). Statistical analyses do not support the view that performance is worse for children who are dissimilar from the LENAⓇ original training set. Whether LENAⓇ results are accurate enough for a given research, educational, or clinical application depends largely on the specifics at hand. We therefore conclude with a set of recommendations to help researchers make this determination for their goals.


Asunto(s)
Lenguaje , Habla , Niño , Lenguaje Infantil , Preescolar , Comunicación , Escolaridad , Humanos , Lactante , Desarrollo del Lenguaje , Masculino
9.
Behav Res Methods ; 53(2): 818-835, 2021 04.
Artículo en Inglés | MEDLINE | ID: mdl-32875399

RESUMEN

Recordings captured by wearable microphones are a standard method for investigating young children's language environments. A key measure to quantify from such data is the amount of speech present in children's home environments. To this end, the LENA recorder and software-a popular system for measuring linguistic input-estimates the number of adult words that children may hear over the course of a recording. However, word count estimation is challenging to do in a language- independent manner; the relationship between observable acoustic patterns and language-specific lexical entities is far from uniform across human languages. In this paper, we ask whether some alternative linguistic units, namely phone(me)s or syllables, could be measured instead of, or in parallel with, words in order to achieve improved cross-linguistic applicability and comparability of an automated system for measuring child language input. We discuss the advantages and disadvantages of measuring different units from theoretical and technical points of view. We also investigate the practical applicability of measuring such units using a novel system called Automatic LInguistic unit Count Estimator (ALICE) together with audio from seven child-centered daylong audio corpora from diverse cultural and linguistic environments. We show that language-independent measurement of phoneme counts is somewhat more accurate than syllables or words, but all three are highly correlated with human annotations on the same data. We share an open-source implementation of ALICE for use by the language research community, enabling automatic phoneme, syllable, and word count estimation from child-centered audio recordings.


Asunto(s)
Lenguaje , Habla , Acústica , Adulto , Niño , Lenguaje Infantil , Preescolar , Humanos , Desarrollo del Lenguaje
10.
Sci Rep ; 10(1): 169, 2020 01 13.
Artículo en Inglés | MEDLINE | ID: mdl-31932616

RESUMEN

Infants' spontaneous and voluntary movements mirror developmental integrity of brain networks since they require coordinated activation of multiple sites in the central nervous system. Accordingly, early detection of infants with atypical motor development holds promise for recognizing those infants who are at risk for a wide range of neurodevelopmental disorders (e.g., cerebral palsy, autism spectrum disorders). Previously, novel wearable technology has shown promise for offering efficient, scalable and automated methods for movement assessment in adults. Here, we describe the development of an infant wearable, a multi-sensor smart jumpsuit that allows mobile accelerometer and gyroscope data collection during movements. Using this suit, we first recorded play sessions of 22 typically developing infants of approximately 7 months of age. These data were manually annotated for infant posture and movement based on video recordings of the sessions, and using a novel annotation scheme specifically designed to assess the overall movement pattern of infants in the given age group. A machine learning algorithm, based on deep convolutional neural networks (CNNs) was then trained for automatic detection of posture and movement classes using the data and annotations. Our experiments show that the setup can be used for quantitative tracking of infant movement activities with a human equivalent accuracy, i.e., it meets the human inter-rater agreement levels in infant posture and movement classification. We also quantify the ambiguity of human observers in analyzing infant movements, and propose a method for utilizing this uncertainty for performance improvements in training of the automated classifier. Comparison of different sensor configurations also shows that four-limb recording leads to the best performance in posture and movement classification.


Asunto(s)
Algoritmos , Cinestesia/fisiología , Monitoreo Ambulatorio/instrumentación , Movimiento/fisiología , Postura/fisiología , Dispositivos Electrónicos Vestibles/estadística & datos numéricos , Automatización , Femenino , Humanos , Lactante , Masculino , Redes Neurales de la Computación , Grabación en Video
11.
Cognition ; 178: 193-206, 2018 09.
Artículo en Inglés | MEDLINE | ID: mdl-29885600

RESUMEN

The exaggerated intonation and special rhythmic properties of infant-directed speech (IDS) have been hypothesized to attract infants' attention to the speech stream. However, there has been little work actually connecting the properties of IDS to models of attentional processing or perceptual learning. A number of such attention models suggest that surprising or novel perceptual inputs attract attention, where novelty can be operationalized as the statistical (un)predictability of the stimulus in the given context. Since prosodic patterns such as F0 contours are accessible to young infants who are also known to be adept statistical learners, the present paper investigates a hypothesis that F0 contours in IDS are less predictable than those in adult-directed speech (ADS), given previous exposure to both speaking styles, thereby potentially tapping into basic attentional mechanisms of the listeners in a similar manner that relative probabilities of other linguistic patterns are known to modulate attentional processing in infants and adults. Computational modeling analyses with naturalistic IDS and ADS speech from matched speakers and contexts show that IDS intonation has lower overall temporal predictability even when the F0 contours of both speaking styles are normalized to have equal means and variances. A closer analysis reveals that there is a tendency of IDS intonation to be less predictable at the end of short utterances, whereas ADS exhibits more stable average predictability patterns across the full extent of the utterances. The difference between IDS and ADS persists even when the proportion of IDS and ADS exposure is varied substantially, simulating different relative amounts of IDS heard in different family and cultural environments. Exposure to IDS is also found to be more efficient for predicting ADS intonation contours in new utterances than exposure to the equal amount of ADS speech. This indicates that the more variable prosodic contours of IDS also generalize to ADS, and may therefore enhance prosodic learning in infancy. Overall, the study suggests that one reason behind infant preference for IDS could be its higher information value at the prosodic level, as measured by the amount of surprisal in the F0 contours. This provides the first formal link between the properties of IDS and the models of attentional processing and statistical learning in the brain. However, this finding does not rule out the possibility that other differences between the IDS and ADS also play a role.


Asunto(s)
Atención , Aprendizaje , Percepción del Habla , Habla , Estimulación Acústica , Humanos , Lactante , Desarrollo del Lenguaje , Fonética , Acústica del Lenguaje
12.
Neuropsychologia ; 109: 181-199, 2018 01 31.
Artículo en Inglés | MEDLINE | ID: mdl-29247667

RESUMEN

Perceptual prominence of linguistic units such as words has been earlier connected to the concepts of predictability and attentional orientation. One hypothesis is that low-probability prosodic or lexical content is perceived as prominent due to the surprisal and high information value associated with the stimulus. However, the existing behavioral studies have used stimulus manipulations that follow or violate typical linguistic patterns present in the listeners' native language, i.e., assuming that the listeners have already established a model for acceptable prosodic patterns in the language. In the present study, we investigated whether prosodic expectations and the resulting subjective impression of prominence is affected by brief statistical adaptation to suprasegmental acoustic features in speech, also in the case where the prosodic patterns do not necessarily follow language-typical marking for prominence. We first exposed listeners to five minutes of speech with uneven distributions of falling and rising fundamental frequency (F0) trajectories on sentence-final words, and then tested their judgments of prominence on a set of new utterances. The results show that the probability of the F0 trajectory affects the perception of prominence, a less frequent F0 trajectory making a word more prominent independently of the absolute direction of F0 change. In the second part of the study, we conducted EEG-measurements on a set of new subjects listening to similar utterances with predominantly rising or falling F0 on sentence-final words. Analysis of the resulting event-related potentials (ERP) reveals a significant difference in N200 and N400 ERP-component amplitudes between standard and deviant prosody, again independently of the F0 direction and the underlying lexical content. Since N400 has earlier been associated with semantic processing of stimuli, this suggests that listeners implicitly track probabilities at the suprasegmental level and that predictability of a prosodic pattern during a word has an impact to the semantic processing of the word. Overall, the study suggests that prosodic markers for prominence are at least partially driven by the statistical structure of recently perceived speech, and therefore prominence perception could be based on statistical learning mechanisms similar to those observed in early word learning, but in this case operating at the level of suprasegmental acoustic features.


Asunto(s)
Anticipación Psicológica/fisiología , Encéfalo/fisiología , Psicolingüística , Semántica , Acústica del Lenguaje , Percepción del Habla/fisiología , Adulto , Electroencefalografía , Potenciales Evocados , Femenino , Humanos , Juicio/fisiología , Masculino , Modelos Psicológicos , Aprendizaje por Probabilidad
13.
Cognition ; 171: 130-150, 2018 02.
Artículo en Inglés | MEDLINE | ID: mdl-29156241

RESUMEN

Syllables are often considered to be central to infant and adult speech perception. Many theories and behavioral studies on early language acquisition are also based on syllable-level representations of spoken language. There is little clarity, however, on what sort of pre-linguistic "syllable" would actually be accessible to an infant with no phonological or lexical knowledge. Anchored by the notion that syllables are organized around particularly sonorous (audible) speech sounds, the present study investigates the feasibility of speech segmentation into syllable-like chunks without any a priori linguistic knowledge. We first operationalize sonority as a measurable property of the acoustic input, and then use sonority variation across time, or speech rhythm, as the basis for segmentation. The entire process from acoustic input to chunks of syllable-like acoustic segments is implemented as a computational model inspired by the oscillatory entrainment of the brain to speech rhythm. We analyze the output of the segmentation process in three different languages, showing that the sonority fluctuation in speech is highly informative of syllable and word boundaries in all three cases without any language-specific tuning of the model. These findings support the widely held assumption that syllable-like structure is accessible to infants even when they are only beginning to learn the properties of their native language.


Asunto(s)
Desarrollo del Lenguaje , Modelos Teóricos , Psicolingüística , Percepción del Habla/fisiología , Humanos
14.
Cogn Sci ; 40(7): 1739-1774, 2016 09.
Artículo en Inglés | MEDLINE | ID: mdl-26481111

RESUMEN

Numerous studies have examined the acoustic correlates of sentential stress and its underlying linguistic functionality. However, the mechanism that connects stress cues to the listener's attentional processing has remained unclear. Also, the learnability versus innateness of stress perception has not been widely discussed. In this work, we introduce a novel perspective to the study of sentential stress and put forward the hypothesis that perceived sentence stress in speech is related to the unpredictability of prosodic features, thereby capturing the attention of the listener. As predictability is based on the statistical structure of the speech input, the hypothesis also suggests that stress perception is a result of general statistical learning mechanisms. To study this idea, computational simulations are performed where temporal prosodic trajectories are modeled with an n-gram model. Probabilities of the feature trajectories are subsequently evaluated on a set of novel utterances and compared to human perception of stress. The results show that the low-probability regions of F0 and energy trajectories are strongly correlated with stress perception, giving support to the idea that attention and unpredictability of sensory stimulus are mutually connected.


Asunto(s)
Atención/fisiología , Aprendizaje/fisiología , Percepción del Habla/fisiología , Adulto , Femenino , Humanos , Lenguaje , Masculino , Persona de Mediana Edad , Modelos Teóricos , Adulto Joven
15.
IEEE Trans Neural Netw Learn Syst ; 27(9): 1878-89, 2016 09.
Artículo en Inglés | MEDLINE | ID: mdl-26285224

RESUMEN

Modeling and prediction of temporal sequences is central to many signal processing and machine learning applications. Prediction based on sequence history is typically performed using parametric models, such as fixed-order Markov chains ( n -grams), approximations of high-order Markov processes, such as mixed-order Markov models or mixtures of lagged bigram models, or with other machine learning techniques. This paper presents a method for sequence prediction based on sparse hyperdimensional coding of the sequence structure and describes how higher order temporal structures can be utilized in sparse coding in a balanced manner. The method is purely incremental, allowing real-time online learning and prediction with limited computational resources. Experiments with prediction of mobile phone use patterns, including the prediction of the next launched application, the next GPS location of the user, and the next artist played with the phone media player, reveal that the proposed method is able to capture the relevant variable-order structure from the sequences. In comparison with the n -grams and the mixed-order Markov models, the sparse hyperdimensional predictor clearly outperforms its peers in terms of unweighted average recall and achieves an equal level of weighted average recall as the mixed-order Markov chain but without the batch training of the mixed-order model.

16.
Cogsci ; 2016: 1757-1762, 2016 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-29359204

RESUMEN

Infants' speech perception adapts to the phonemic categories of their native language, a process assumed to be driven by the distributional properties of speech. This study investigates whether deep neural networks (DNNs), the current state-of-the-art in distributional feature learning, are capable of learning phoneme-like representations of speech in an unsupervised manner. We trained DNNs with unlabeled and labeled speech and analyzed the activations of each layer with respect to the phones in the input segments. The analyses reveal that the emergence of phonemic invariance in DNNs is dependent on the availability of phonemic labeling of the input during the training. No increased phonemic selectivity of the hidden layers was observed in the purely unsupervised networks despite successful learning of low-dimensional representations for speech. This suggests that additional learning constraints or more sophisticated models are needed to account for the emergence of phone-like categories in distributional learning operating on natural speech.

17.
Psychol Rev ; 122(4): 792-829, 2015 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-26437151

RESUMEN

Human infants learn meanings for spoken words in complex interactions with other people, but the exact learning mechanisms are unknown. Among researchers, a widely studied learning mechanism is called cross-situational learning (XSL). In XSL, word meanings are learned when learners accumulate statistical information between spoken words and co-occurring objects or events, allowing the learner to overcome referential uncertainty after having sufficient experience with individually ambiguous scenarios. Existing models in this area have mainly assumed that the learner is capable of segmenting words from speech before grounding them to their referential meaning, while segmentation itself has been treated relatively independently of the meaning acquisition. In this article, we argue that XSL is not just a mechanism for word-to-meaning mapping, but that it provides strong cues for proto-lexical word segmentation. If a learner directly solves the correspondence problem between continuous speech input and the contextual referents being talked about, segmentation of the input into word-like units emerges as a by-product of the learning. We present a theoretical model for joint acquisition of proto-lexical segments and their meanings without assuming a priori knowledge of the language. We also investigate the behavior of the model using a computational implementation, making use of transition probability-based statistical learning. Results from simulations show that the model is not only capable of replicating behavioral data on word learning in artificial languages, but also shows effective learning of word segments and their meanings from continuous speech. Moreover, when augmented with a simple familiarity preference during learning, the model shows a good fit to human behavioral data in XSL tasks. These results support the idea of simultaneous segmentation and meaning acquisition and show that comprehensive models of early word segmentation should take referential word meanings into account. (PsycINFO Database Record


Asunto(s)
Desarrollo del Lenguaje , Lenguaje , Aprendizaje/fisiología , Modelos Psicológicos , Humanos
18.
Annu Int Conf IEEE Eng Med Biol Soc ; 2015: 1492-5, 2015 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-26736553

RESUMEN

Essential information about early brain maturation can be retrieved from the preterm human electroencephalogram (EEG). This study proposes a new set of quantitative features that correlate with early maturation. We exploit the known early trend in EEG content from intermittent to continuous activity, which changes the line length content of the EEG. The developmental shift can be captured in the line length histogram, which we use to obtain 28 features; 20 histogram bins and 8 other statistical measurements. Using the mutual information, we select 6 features with high correlation to the infant's age. This subset appears promising to detect deviances from normal brain maturation. The presented data-driven index holds promise for developing into a computational EEG index of maturation that is highly needed for overall assessment in the Neonatal Intensive Care Units.


Asunto(s)
Electroencefalografía , Encéfalo , Humanos , Recién Nacido , Recien Nacido Prematuro , Unidades de Cuidado Intensivo Neonatal , Conducta Social
19.
Front Hum Neurosci ; 8: 1030, 2014.
Artículo en Inglés | MEDLINE | ID: mdl-25566040

RESUMEN

A key feature of normal neonatal EEG at term age is interhemispheric synchrony (IHS), which refers to the temporal co-incidence of bursting across hemispheres during trace alternant EEG activity. The assessment of IHS in both clinical and scientific work relies on visual, qualitative EEG assessment without clearly quantifiable definitions. A quantitative measure, activation synchrony index (ASI), was recently shown to perform well as compared to visual assessments. The present study was set out to test whether IHS is stable enough for clinical use, and whether it could be an objective feature of EEG normality. We analyzed 31 neonatal EEG recordings that had been clinically classified as normal (n = 14) or abnormal (n = 17) using holistic, conventional visual criteria including amplitude, focal differences, qualitative synchrony, and focal abnormalities. We selected 20-min epochs of discontinuous background pattern. ASI values were computed separately for different channel pair combinations and window lengths to define them for the optimal ASI intraindividual stability. Finally, ROC curves were computed to find trade-offs related to compromised data lengths, a common challenge in neonatal EEG studies. Using the average of four consecutive 2.5-min epochs in the centro-occipital bipolar derivations gave ASI estimates that very accurately distinguished babies clinically classified as normal vs. abnormal. It was even possible to draw a cut-off limit (ASI~3.6) which correctly classified the EEGs in 97% of all cases. Finally, we showed that compromising the length of EEG segments from 20 to 5 min leads to increased variability in ASI-based classification. Our findings support the prior literature that IHS is an important feature of normal neonatal brain function. We show that ASI may provide diagnostic value even at individual level, which strongly supports its use in prospective clinical studies on neonatal EEG as well as in the feature set of upcoming EEG classifiers.

20.
J Acoust Soc Am ; 134(1): 407-19, 2013 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-23862817

RESUMEN

Several psychoacoustic phenomena such as loudness perception, absolute thresholds of hearing, and perceptual grouping in time are affected by temporal integration of the signal in the auditory system. Similarly, the frequency resolution of the hearing system, often expressed in terms of critical bands, implies signal integration across neighboring frequencies. Although progress has been made in understanding the neurophysiological mechanisms behind these processes, the underlying reasons for the observed integration characteristics have remained poorly understood. The current work proposes that the temporal and spectral integration are a result of a system optimized for pattern detection from ecologically relevant acoustic inputs. This argument is supported by a simulation where the average time-frequency structure of speech that is derived from a large set of speech signals shows a good match to the time-frequency characteristics of the human auditory system. The results also suggest that the observed integration characteristics are learnable from acoustic inputs of the auditory environment using a Hebbian-like learning rule.


Asunto(s)
Enmascaramiento Perceptual , Discriminación de la Altura Tonal , Espectrografía del Sonido , Acústica del Lenguaje , Percepción del Habla , Percepción del Tiempo , Atención , Señales (Psicología) , Humanos
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA
...