Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 20
Filtrar
1.
Proc Natl Acad Sci U S A ; 115(15): 3972-3977, 2018 04 10.
Artículo en Inglés | MEDLINE | ID: mdl-29581266

RESUMEN

Human listeners excel at forming high-level social representations about each other, even from the briefest of utterances. In particular, pitch is widely recognized as the auditory dimension that conveys most of the information about a speaker's traits, emotional states, and attitudes. While past research has primarily looked at the influence of mean pitch, almost nothing is known about how intonation patterns, i.e., finely tuned pitch trajectories around the mean, may determine social judgments in speech. Here, we introduce an experimental paradigm that combines state-of-the-art voice transformation algorithms with psychophysical reverse correlation and show that two of the most important dimensions of social judgments, a speaker's perceived dominance and trustworthiness, are driven by robust and distinguishing pitch trajectories in short utterances like the word "Hello," which remained remarkably stable whether male or female listeners judged male or female speakers. These findings reveal a unique communicative adaptation that enables listeners to infer social traits regardless of speakers' physical characteristics, such as sex and mean pitch. By characterizing how any given individual's mental representations may differ from this generic code, the method introduced here opens avenues to explore dysprosody and social-cognitive deficits in disorders like autism spectrum and schizophrenia. In addition, once derived experimentally, these prototypes can be applied to novel utterances, thus providing a principled way to modulate personality impressions in arbitrary speech signals.


Asunto(s)
Conducta Social , Habla , Voz , Adulto , Percepción Auditiva , Emociones , Femenino , Humanos , Juicio , Masculino , Persona de Mediana Edad , Distribución Aleatoria , Percepción del Habla , Confianza , Adulto Joven
2.
J Acoust Soc Am ; 150(5): 3631, 2021 11.
Artículo en Inglés | MEDLINE | ID: mdl-34852611

RESUMEN

Amplitude modulation (AM) and frequency modulation (FM) provide crucial auditory information. If FM is encoded as AM, it should be possible to give a unified account of AM and FM perception both in terms of response consistency and performance. These two aspects of behavior were estimated for normal-hearing participants using a constant-stimuli, forced-choice detection task repeated twice with the same stimuli (double pass). Sinusoidal AM or FM with rates of 2 or 20 Hz were applied to a 500-Hz pure-tone carrier and presented at detection threshold. All stimuli were masked by a modulation noise. Percent agreement of responses across passes and percent-correct detection for the two passes were used to estimate consistency and performance, respectively. These data were simulated using a model implementing peripheral processes, a central modulation filterbank, an additive internal noise, and a template-matching device. Different levels of internal noise were required to reproduce AM and FM data, but a single level could account for the 2- and 20-Hz AM data. As for FM, two levels of internal noise were needed to account for detection at slow and fast rates. Finally, the level of internal noise yielding best predictions increased with the level of the modulation-noise masker. Overall, these results suggest that different sources of internal variability are involved for AM and FM detection at low audio frequencies.


Asunto(s)
Percepción Auditiva , Ruido , Umbral Auditivo , Audición , Humanos , Ruido/efectos adversos
3.
J Acoust Soc Am ; 143(1): EL19, 2018 01.
Artículo en Inglés | MEDLINE | ID: mdl-29390775

RESUMEN

Which spectral cues underlie the perceptual processing of smiles in speech? Here, the question was addressed using reverse-correlation in the case of the isolated vowel [a]. Listeners were presented with hundreds of pairs of utterances with randomly manipulated spectral characteristics and were asked to indicate, in each pair, which was the most smiling. The analyses revealed that they relied on robust spectral representations that specifically encoded vowel's formants. These findings demonstrate the causal role played by formants in the perception of smile. Overall, this paper suggests a general method to estimate the spectral bases of high-level (e.g., emotional/social/paralinguistic) speech representations.


Asunto(s)
Señales (Psicología) , Sonrisa , Acústica del Lenguaje , Percepción del Habla , Calidad de la Voz , Estimulación Acústica , Adolescente , Adulto , Femenino , Humanos , Masculino , Espectrografía del Sonido , Adulto Joven
4.
J Acoust Soc Am ; 142(1): 256, 2017 07.
Artículo en Inglés | MEDLINE | ID: mdl-28764470

RESUMEN

The mechanisms underlying global loudness judgments of rising- or falling-intensity tones were further investigated in two magnitude estimation experiments. By manipulating the temporal characteristics of such stimuli, it was examined whether judgments could be accounted for by an integration of their loudest portion over a certain temporal window associated to a "decay mechanism" downsizing this integration over time for falling ramps. In experiment 1, 1-kHz intensity-ramps were stretched in time between 1 and 16 s keeping their dynamics (difference between maximum and minimum levels) unchanged. While global loudness of rising tones increased up to 6 s, evaluations of falling tones increased at a weaker rate and slightly decayed between 6 and 16 s, resulting in significant differences between the two patterns. In experiment 2, ramps were stretched in time between 2 and 12 s keeping their slopes (rate of change in dB/s) unchanged. In this context, the main effect of duration became non-significant and the interaction between the two profiles remained, although the decay of falling tones was not significant. These results qualitatively support the view that the global loudness computation of intensity-ramps involves an integration of their loudest portions; the presence of a decay mechanism could, however, not be attested.

7.
J Acoust Soc Am ; 139(1): 406-17, 2016 Jan.
Artículo en Inglés | MEDLINE | ID: mdl-26827035

RESUMEN

Psychophysical studies on loudness have so far examined the temporal weighting of loudness solely in level-discrimination tasks. Typically, listeners were asked to discriminate hundreds of level-fluctuating sounds regarding their global loudness. Temporal weights, i.e., the importance of each temporal portion of the stimuli for the loudness judgment, were then estimated from listeners' responses. Consistent non-uniform "u-shaped" temporal weighting patterns were observed, with greater weights assigned to the first and the last temporal portions of the stimuli, revealing significant primacy and recency effects, respectively. In this study, the question was addressed whether the same weighting pattern could be found in a traditional loudness estimation task. Temporal loudness weights were compared between a level-discrimination (LD) task and an absolute magnitude estimation (AME) task. Stimuli were 3-s broadband noises consisting of 250-ms segments randomly varying in level. Listeners were asked to evaluate the global loudness of the stimuli by classifying them as "loud" or "soft" (LD), or by assigning a number representing their loudness (AME). Results showed non-uniform temporal weighting in both tasks, but also significant differences between the two tasks. An explanation based on the difference in complexity between the evaluation processes underlying each task is proposed.


Asunto(s)
Percepción Sonora/fisiología , Estimulación Acústica , Adulto , Análisis de Varianza , Toma de Decisiones/fisiología , Femenino , Humanos , Juicio/fisiología , Masculino , Ruido , Enmascaramiento Perceptual/fisiología , Pruebas Psicológicas , Adulto Joven
8.
Sci Rep ; 14(1): 15194, 2024 07 02.
Artículo en Inglés | MEDLINE | ID: mdl-38956187

RESUMEN

After a right hemisphere stroke, more than half of the patients are impaired in their capacity to produce or comprehend speech prosody. Yet, and despite its social-cognitive consequences for patients, aprosodia following stroke has received scant attention. In this report, we introduce a novel, simple psychophysical procedure which, by combining systematic digital manipulations of speech stimuli and reverse-correlation analysis, allows estimating the internal sensory representations that subtend how individual patients perceive speech prosody, and the level of internal noise that govern behavioral variability in how patients apply these representations. Tested on a sample of N = 22 right-hemisphere stroke survivors and N = 21 age-matched controls, the representation + noise model provides a promising alternative to the clinical gold standard for evaluating aprosodia (MEC): both parameters strongly associate with receptive, and not expressive, aprosodia measured by MEC within the patient group; they have better sensitivity than MEC for separating high-functioning patients from controls; and have good specificity with respect to non-prosody-related impairments of auditory attention and processing. Taken together, individual differences in either internal representation, internal noise, or both, paint a potent portrait of the variety of sensory/cognitive mechanisms that can explain impairments of prosody processing after stroke.


Asunto(s)
Percepción del Habla , Accidente Cerebrovascular , Humanos , Accidente Cerebrovascular/fisiopatología , Accidente Cerebrovascular/complicaciones , Percepción del Habla/fisiología , Masculino , Femenino , Persona de Mediana Edad , Anciano , Ruido , Psicofísica/métodos , Adulto
9.
J Acoust Soc Am ; 134(4): EL321-6, 2013 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-24116537

RESUMEN

Using molecular psychophysics, temporal loudness weights were measured for 2-s, 1-kHz tones with flat, increasing and decreasing time-intensity profiles. While primacy and recency effects were observed for flat profile stimuli, the so-called "level dominance" effect was observed for both increasing and decreasing profile stimuli, fully determining their temporal weights. The weighs obtained for these profiles were basically zero for all but the most intense parts of these sounds. This supports the view that the "level dominance" effect is prominent with intensity-varying sounds and that it persists over time since temporal weights are not affected by the direction of intensity change.


Asunto(s)
Percepción Sonora , Percepción del Tiempo , Estimulación Acústica , Adulto , Audiometría de Tonos Puros , Umbral Auditivo , Femenino , Lateralidad Funcional , Humanos , Masculino , Psicoacústica , Factores de Tiempo , Adulto Joven
10.
JASA Express Lett ; 3(8)2023 08 01.
Artículo en Inglés | MEDLINE | ID: mdl-37566904

RESUMEN

Temporal and frequency auditory streaming capacities were assessed for non-musician (NM), expert musician (EM), and amateur musician (AM) listeners using a local-global task and an interleaved melody recognition task, respectively. Data replicate differences previously observed between NM and EM, and reveal that while AM exhibits a local-over-global processing change comparable to EM, their performance for segregating a melody embedded in a stream remains as poor as NM. The observed group partitioning along the temporal-frequency auditory streaming capacity map suggests a sequential, two-step development model of musical learning, whose contributing factors are discussed.


Asunto(s)
Música , Reconocimiento en Psicología
11.
Autism ; 27(3): 629-646, 2023 04.
Artículo en Inglés | MEDLINE | ID: mdl-35848413

RESUMEN

LAY ABSTRACT: As a key auditory attribute of sounds, pitch is ubiquitous in our everyday listening experience involving language, music and environmental sounds. Given its critical role in auditory processing related to communication, numerous studies have investigated pitch processing in autism spectrum disorder. However, the findings have been mixed, reporting either enhanced, typical or impaired performance among autistic individuals. By investigating top-down comparisons of internal mental representations of pitch contours in speech and music, this study shows for the first time that, while autistic individuals exhibit diverse profiles of pitch processing compared to non-autistic individuals, their mental representations of pitch contours are typical across domains. These findings suggest that pitch-processing mechanisms are shared across domains in autism spectrum disorder and provide theoretical implications for using music to improve speech for those autistic individuals who have language problems.


Asunto(s)
Trastorno del Espectro Autista , Música , Percepción del Habla , Humanos , Percepción de la Altura Tonal , Habla
12.
Nat Commun ; 12(1): 861, 2021 02 08.
Artículo en Inglés | MEDLINE | ID: mdl-33558510

RESUMEN

The success of human cooperation crucially depends on mechanisms enabling individuals to detect unreliability in their conspecifics. Yet, how such epistemic vigilance is achieved from naturalistic sensory inputs remains unclear. Here we show that listeners' perceptions of the certainty and honesty of other speakers from their speech are based on a common prosodic signature. Using a data-driven method, we separately decode the prosodic features driving listeners' perceptions of a speaker's certainty and honesty across pitch, duration and loudness. We find that these two kinds of judgments rely on a common prosodic signature that is perceived independently from individuals' conceptual knowledge and native language. Finally, we show that listeners extract this prosodic signature automatically, and that this impacts the way they memorize spoken words. These findings shed light on a unique auditory adaptation that enables human listeners to quickly detect and react to unreliability during linguistic interactions.


Asunto(s)
Percepción del Habla/fisiología , Habla/fisiología , Conducta de Elección , Toma de Decisiones , Femenino , Humanos , Juicio , Conocimiento , Lenguaje , Lingüística , Masculino , Memoria a Corto Plazo , Análisis y Desempeño de Tareas , Adulto Joven
13.
Trends Hear ; 25: 2331216520978029, 2021.
Artículo en Inglés | MEDLINE | ID: mdl-33620023

RESUMEN

Spectrotemporal modulations (STM) are essential features of speech signals that make them intelligible. While their encoding has been widely investigated in neurophysiology, we still lack a full understanding of how STMs are processed at the behavioral level and how cochlear hearing loss impacts this processing. Here, we introduce a novel methodological framework based on psychophysical reverse correlation deployed in the modulation space to characterize the mechanisms underlying STM detection in noise. We derive perceptual filters for young normal-hearing and older hearing-impaired individuals performing a detection task of an elementary target STM (a given product of temporal and spectral modulations) embedded in other masking STMs. Analyzed with computational tools, our data show that both groups rely on a comparable linear (band-pass)-nonlinear processing cascade, which can be well accounted for by a temporal modulation filter bank model combined with cross-correlation against the target representation. Our results also suggest that the modulation mistuning observed for the hearing-impaired group results primarily from broader cochlear filters. Yet, we find idiosyncratic behaviors that cannot be captured by cochlear tuning alone, highlighting the need to consider variability originating from additional mechanisms. Overall, this integrated experimental-computational approach offers a principled way to assess suprathreshold processing distortions in each individual and could thus be used to further investigate interindividual differences in speech intelligibility.


Asunto(s)
Pérdida Auditiva Sensorineural , Percepción del Habla , Umbral Auditivo , Audición , Pérdida Auditiva Sensorineural/diagnóstico , Humanos , Ruido/efectos adversos , Enmascaramiento Perceptual
14.
Behav Processes ; 172: 104042, 2020 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-31926279

RESUMEN

Many animal vocalizations contain nonlinear acoustic phenomena as a consequence of physiological arousal. In humans, nonlinear features are processed early in the auditory system, and are used to efficiently detect alarm calls and other urgent signals. Yet, high-level emotional and semantic contextual factors likely guide the perception and evaluation of roughness features in vocal sounds. Here we examined the relationship between perceived vocal arousal and auditory context. We presented listeners with nonverbal vocalizations (yells of a single vowel) at varying levels of portrayed vocal arousal, in two musical contexts (clean guitar, distorted guitar) and one non-musical context (modulated noise). As predicted, vocalizations with higher levels of portrayed vocal arousal were judged as more negative and more emotionally aroused than the same voices produced with low vocal arousal. Moreover, both the perceived valence and emotional arousal of vocalizations were significantly affected by both musical and non-musical contexts. These results show the importance of auditory context in judging emotional arousal and valence in voices and music, and suggest that nonlinear features in music are processed similarly to communicative vocal signals.


Asunto(s)
Percepción Auditiva , Emociones , Música , Voz , Adulto , Animales , Nivel de Alerta , Femenino , Humanos , Lenguaje , Masculino , Ruido , Adulto Joven
15.
Sci Rep ; 10(1): 16390, 2020 10 02.
Artículo en Inglés | MEDLINE | ID: mdl-33009439

RESUMEN

The way the visual system processes different scales of spatial information has been widely studied, highlighting the dominant role of global over local processing. Recent studies addressing how the auditory system deals with local-global temporal information suggest a comparable processing scheme, but little is known about how this organization is modulated by long-term musical training, in particular regarding musical sequences. Here, we investigate how non-musicians and expert musicians detect local and global pitch changes in short hierarchical tone sequences structured across temporally-segregated triplets made of musical intervals (local scale) forming a melodic contour (global scale) varying either in one direction (monotonic) or both (non-monotonic). Our data reveal a clearly distinct organization between both groups. Non-musicians show global advantage (enhanced performance to detect global over local modifications) and global-to-local interference effects (interference of global over local processing) only for monotonic sequences, while musicians exhibit the reversed pattern for non-monotonic sequences. These results suggest that the local-global processing scheme depends on the complexity of the melodic contour, and that long-term musical training induces a prominent perceptual reorganization that reshapes its initial global dominance to favour local information processing. This latter result supports the theory of "analytic" processing acquisition in musicians.


Asunto(s)
Percepción Auditiva/fisiología , Discriminación de la Altura Tonal/fisiología , Estimulación Acústica/métodos , Adulto , Cognición/fisiología , Potenciales Evocados Auditivos/fisiología , Femenino , Humanos , Masculino , Música , Percepción de la Altura Tonal/fisiología , Tiempo de Reacción/fisiología , Percepción del Tiempo/fisiología , Adulto Joven
16.
PLoS One ; 14(4): e0205943, 2019.
Artículo en Inglés | MEDLINE | ID: mdl-30947281

RESUMEN

Over the past few years, the field of visual social cognition and face processing has been dramatically impacted by a series of data-driven studies employing computer-graphics tools to synthesize arbitrary meaningful facial expressions. In the auditory modality, reverse correlation is traditionally used to characterize sensory processing at the level of spectral or spectro-temporal stimulus properties, but not higher-level cognitive processing of e.g. words, sentences or music, by lack of tools able to manipulate the stimulus dimensions that are relevant for these processes. Here, we present an open-source audio-transformation toolbox, called CLEESE, able to systematically randomize the prosody/melody of existing speech and music recordings. CLEESE works by cutting recordings in small successive time segments (e.g. every successive 100 milliseconds in a spoken utterance), and applying a random parametric transformation of each segment's pitch, duration or amplitude, using a new Python-language implementation of the phase-vocoder digital audio technique. We present here two applications of the tool to generate stimuli for studying intonation processing of interrogative vs declarative speech, and rhythm processing of sung melodies.


Asunto(s)
Cognición/fisiología , Música , Percepción del Habla/fisiología , Habla/fisiología , Femenino , Humanos , Masculino
18.
Nat Commun ; 7: 12682, 2016 09 01.
Artículo en Inglés | MEDLINE | ID: mdl-27580932

RESUMEN

Sound recognition relies not only on spectral cues, but also on temporal cues, as demonstrated by the profound impact of time reversals on perception of common sounds. To address the coding principles underlying such auditory asymmetries, we recorded a large sample of auditory cortex neurons using two-photon calcium imaging in awake mice, while playing sounds ramping up or down in intensity. We observed clear asymmetries in cortical population responses, including stronger cortical activity for up-ramping sounds, which matches perceptual saliency assessments in mice and previous measures in humans. Analysis of cortical activity patterns revealed that auditory cortex implements a map of spatially clustered neuronal ensembles, detecting specific combinations of spectral and intensity modulation features. Comparing different models, we show that cortical responses result from multi-layered nonlinearities, which, contrary to standard receptive field models of auditory cortex function, build divergent representations of sounds with similar spectral content, but different temporal structure.


Asunto(s)
Estimulación Acústica , Corteza Auditiva/fisiología , Percepción Auditiva/fisiología , Nervio Coclear/fisiología , Audición/fisiología , Animales , Nervio Coclear/citología , Masculino , Ratones , Ratones Endogámicos C57BL , Microscopía Confocal/métodos , Neuronas/fisiología , Sonido
19.
Atten Percept Psychophys ; 77(3): 907-20, 2015 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-25609028

RESUMEN

Tones rising in intensity over a few seconds are perceived as louder than symmetrical tones falling in intensity. However, the causes for such perceptual asymmetry, as well as its magnitude and dependency on contextual and methodological factors remain unclear. In this paper, two psychophysical experiments were conducted to measure the magnitude of this asymmetry for 2-s, 15-dB intensity-varying tones in different conditions. In the first experiment, participants assessed the global loudness of rising- and falling-intensity sounds with an absolute magnitude estimation procedure (AME); in the second experiment, they compared sounds relatively in an adaptive, two-interval, two-alternative forced-choice task (2I-2AFC). In both experiments, the region of intensity change, the design of experimental blocks, and the type of comparison stimulus were systematically manipulated to test for contextual and methodological factors. Remarkably, the asymmetry was virtually unaffected by the different contexts of presentation and similar results with 2I-2AFC and AME measurements were obtained. In addition, the size of the effect was comparable over all but the highest intensity regions (80-90 dB SPL), at which it was significantly smaller. All together, these results indicate that the loudness asymmetry is preserved under different measurement methods and contexts, and suggest that the underlying mechanism is strong and robust. In short, falling tones have to be about 4 dB higher in level than symmetrically rising tones in order to be perceived with the same global loudness, a finding that is still not predicted by current loudness models.


Asunto(s)
Audición/fisiología , Percepción Sonora/fisiología , Adulto , Análisis de Varianza , Conducta de Elección , Femenino , Humanos , Masculino , Modelos Teóricos , Enmascaramiento Perceptual/fisiología , Psicoacústica , Valores de Referencia , Sonido , Adulto Joven
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA