Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 5.502
Filter
1.
J Acoust Soc Am ; 156(1): 262-277, 2024 Jul 01.
Article in English | MEDLINE | ID: mdl-38980101

ABSTRACT

A series of Bayesian adaptive procedures to estimate loudness growth across a wide frequency range from individual listeners was developed, and these procedures were compared. Simulation experiments were conducted based on multinomial psychometric functions for categorical loudness scaling across ten test frequencies estimated from 61 listeners with normal hearing and 87 listeners with sensorineural hearing loss. Adaptive procedures that optimized the stimulus selection based on the interim estimates of two types of category-boundary models were tested. The first type of model was a phenomenological model of category boundaries adopted from previous research studies, while the other type was a data-driven model derived from a previously collected set of categorical loudness scaling data. An adaptive procedure without Bayesian active learning was also implemented. Results showed that all adaptive procedures provided convergent estimates of the loudness category boundaries and equal-loudness contours between 250 and 8000 Hz. Performing post hoc model fitting, using the data-driven model, on the collected data led to satisfactory accuracies, such that all adaptive procedures tested in the current study, independent of modeling approach and stimulus-selection rules, were able to provide estimates of the equal-loudness-level contours between 20 and 100 phons with root-mean-square errors typically under 6 dB after 100 trials.


Subject(s)
Acoustic Stimulation , Bayes Theorem , Hearing Loss, Sensorineural , Loudness Perception , Humans , Hearing Loss, Sensorineural/psychology , Hearing Loss, Sensorineural/physiopathology , Adult , Middle Aged , Female , Male , Acoustic Stimulation/methods , Aged , Young Adult , Case-Control Studies , Auditory Threshold , Computer Simulation , Psychoacoustics
2.
J Acoust Soc Am ; 156(1): 326-340, 2024 Jul 01.
Article in English | MEDLINE | ID: mdl-38990035

ABSTRACT

Humans are adept at identifying spectral patterns, such as vowels, in different rooms, at different sound levels, or produced by different talkers. How this feat is achieved remains poorly understood. Two psychoacoustic analogs of spectral pattern recognition are spectral profile analysis and spectrotemporal ripple direction discrimination. This study tested whether pattern-recognition abilities observed previously at low frequencies are also observed at extended high frequencies. At low frequencies (center frequency ∼500 Hz), listeners were able to achieve accurate profile-analysis thresholds, consistent with prior literature. However, at extended high frequencies (center frequency ∼10 kHz), listeners' profile-analysis thresholds were either unmeasurable or could not be distinguished from performance based on overall loudness cues. A similar pattern of results was observed with spectral ripple discrimination, where performance was again considerably better at low than at high frequencies. Collectively, these results suggest a severe deficit in listeners' ability to analyze patterns of intensity across frequency in the extended high-frequency region that cannot be accounted for by cochlear frequency selectivity. One interpretation is that the auditory system is not optimized to analyze such fine-grained across-frequency profiles at extended high frequencies, as they are not typically informative for everyday sounds.


Subject(s)
Acoustic Stimulation , Auditory Threshold , Psychoacoustics , Humans , Young Adult , Female , Male , Adult , Cues , Speech Perception/physiology , Sound Spectrography , Loudness Perception , Pattern Recognition, Physiological
3.
Brain Behav ; 14(5): e3520, 2024 May.
Article in English | MEDLINE | ID: mdl-38715412

ABSTRACT

OBJECTIVE: In previous animal studies, sound enhancement reduced tinnitus perception in cases associated with hearing loss. The aim of this study was to investigate the efficacy of sound enrichment therapy in tinnitus treatment by developing a protocol that includes criteria for psychoacoustic characteristics of tinnitus to determine whether the etiology is related to hearing loss. METHODS: A total of 96 patients with chronic tinnitus were included in the study. Fifty-two patients in the study group and 44 patients in the placebo group considered residual inhibition (RI) outcomes and tinnitus pitches. Both groups received sound enrichment treatment with different spectrum contents. The tinnitus handicap inventory (THI), visual analog scale (VAS), minimum masking level (MML), and tinnitus loudness level (TLL) results were compared before and at 1, 3, and 6 months after treatment. RESULTS: There was a statistically significant difference between the groups in THI, VAS, MML, and TLL scores from the first month to all months after treatment (p < .01). For the study group, there was a statistically significant decrease in THI, VAS, MML, and TLL scores in the first month (p < .01). This decrease continued at a statistically significant level in the third month of posttreatment for THI (p < .05) and at all months for VAS-1 (tinnitus severity) (p < .05) and VAS-2 (tinnitus discomfort) (p < .05). CONCLUSION: In clinical practice, after excluding other factors related to the tinnitus etiology, sound enrichment treatment can be effective in tinnitus cases where RI is positive and the tinnitus pitch is matched with a hearing loss between 45 and 55 dB HL in a relatively short period of 1 month.


Subject(s)
Hearing Loss , Tinnitus , Tinnitus/therapy , Humans , Male , Female , Middle Aged , Adult , Hearing Loss/rehabilitation , Hearing Loss/therapy , Treatment Outcome , Aged , Acoustic Stimulation/methods , Sound , Psychoacoustics
4.
Hear Res ; 448: 109020, 2024 Jul.
Article in English | MEDLINE | ID: mdl-38763034

ABSTRACT

Combining cochlear implants with binaural acoustic hearing via preserved hearing in the implanted ear(s) is commonly referred to as combined electric and acoustic stimulation (EAS). EAS fittings can provide patients with significant benefit for speech recognition in complex noise, perceived listening difficulty, and horizontal-plane localization as compared to traditional bimodal hearing conditions with contralateral and monaural acoustic hearing. However, EAS benefit varies across patients and the degree of benefit is not reliably related to the underlying audiogram. Previous research has indicated that EAS benefit for speech recognition in complex listening scenarios and localization is significantly correlated with the patients' binaural cue sensitivity, namely interaural time differences (ITD). In the context of pure tones, interaural phase differences (IPD) and ITD can be understood as two perspectives on the same phenomenon. Through simple mathematical conversion, one can be transformed into the other, illustrating their inherent interrelation for spatial hearing abilities. However, assessing binaural cue sensitivity is not part of a clinical assessment battery as psychophysical tasks are time consuming, require training to achieve performance asymptote, and specialized programming and software all of which render this clinically unfeasible. In this study, we investigated the possibility of using an objective measure of binaural cue sensitivity by the acoustic change complex (ACC) via imposition of an IPD of varying degrees at stimulus midpoint. Ten adult listeners with normal hearing were assessed on tasks of behavioral and objective binaural cue sensitivity for carrier frequencies of 250 and 1000 Hz. Results suggest that 1) ACC amplitude increases with IPD; 2) ACC-based IPD sensitivity for 250 Hz is significantly correlated with behavioral ITD sensitivity; 3) Participants were more sensitive to IPDs at 250 Hz as compared to 1000 Hz. Thus, this objective measure of IPD sensitivity may hold clinical application for pre- and post-operative assessment for individuals meeting candidacy indications for cochlear implantation with low-frequency acoustic hearing preservation as this relatively quick and objective measure may provide clinicians with information identifying patients most likely to derive benefit from EAS technology.


Subject(s)
Acoustic Stimulation , Auditory Threshold , Cochlear Implantation , Cochlear Implants , Cues , Sound Localization , Speech Perception , Humans , Female , Male , Cochlear Implantation/instrumentation , Adult , Middle Aged , Electric Stimulation , Audiometry, Pure-Tone , Persons With Hearing Impairments/psychology , Persons With Hearing Impairments/rehabilitation , Time Factors , Aged , Noise/adverse effects , Perceptual Masking , Young Adult , Hearing , Psychoacoustics
5.
J Acoust Soc Am ; 155(5): 2990-3004, 2024 May 01.
Article in English | MEDLINE | ID: mdl-38717206

ABSTRACT

Speakers can place their prosodic prominence on any locations within a sentence, generating focus prosody for listeners to perceive new information. This study aimed to investigate age-related changes in the bottom-up processing of focus perception in Jianghuai Mandarin by clarifying the perceptual cues and the auditory processing abilities involved in the identification of focus locations. Young, middle-aged, and older speakers of Jianghuai Mandarin completed a focus identification task and an auditory perception task. The results showed that increasing age led to a decrease in listeners' accuracy rate in identifying focus locations, with all participants performing the worst when dynamic pitch cues were inaccessible. Auditory processing abilities did not predict focus perception performance in young and middle-aged listeners but accounted significantly for the variance in older adults' performance. These findings suggest that age-related deteriorations in focus perception can be largely attributed to declined auditory processing of perceptual cues. Poor ability to extract frequency modulation cues may be the most important underlying psychoacoustic factor for older adults' difficulties in perceiving focus prosody in Jianghuai Mandarin. The results contribute to our understanding of the bottom-up mechanisms involved in linguistic prosody processing in aging adults, particularly in tonal languages.


Subject(s)
Aging , Cues , Speech Perception , Humans , Middle Aged , Aged , Male , Female , Aging/psychology , Aging/physiology , Young Adult , Adult , Speech Perception/physiology , Age Factors , Speech Acoustics , Acoustic Stimulation , Pitch Perception , Language , Voice Quality , Psychoacoustics , Audiometry, Speech
6.
J Acoust Soc Am ; 155(4): 2724-2727, 2024 Apr 01.
Article in English | MEDLINE | ID: mdl-38656337

ABSTRACT

The auditory sensitivity of a small songbird, the red-cheeked cordon bleu, was measured using the standard methods of animal psychophysics. Hearing in cordon bleus is similar to other small passerines with best hearing in the frequency region from 2 to 4 kHz and sensitivity declining at the rate of about 10 dB/octave below 2 kHz and about 35 dB/octave as frequency increases from 4 to 9 kHz. While critical ratios are similar to other songbirds, the long-term average power spectrum of cordon bleu song falls above the frequency of best hearing in this species.


Subject(s)
Acoustic Stimulation , Auditory Threshold , Hearing , Songbirds , Vocalization, Animal , Animals , Vocalization, Animal/physiology , Hearing/physiology , Songbirds/physiology , Male , Psychoacoustics , Sound Spectrography , Female
7.
Sci Rep ; 14(1): 8181, 2024 04 08.
Article in English | MEDLINE | ID: mdl-38589483

ABSTRACT

Temporal envelope modulations (TEMs) are one of the most important features that cochlear implant (CI) users rely on to understand speech. Electroencephalographic assessment of TEM encoding could help clinicians to predict speech recognition more objectively, even in patients unable to provide active feedback. The acoustic change complex (ACC) and the auditory steady-state response (ASSR) evoked by low-frequency amplitude-modulated pulse trains can be used to assess TEM encoding with electrical stimulation of individual CI electrodes. In this study, we focused on amplitude modulation detection (AMD) and amplitude modulation frequency discrimination (AMFD) with stimulation of a basal versus an apical electrode. In twelve adult CI users, we (a) assessed behavioral AMFD thresholds and (b) recorded cortical auditory evoked potentials (CAEPs), AMD-ACC, AMFD-ACC, and ASSR in a combined 3-stimulus paradigm. We found that the electrophysiological responses were significantly higher for apical than for basal stimulation. Peak amplitudes of AMFD-ACC were small and (therefore) did not correlate with speech-in-noise recognition. We found significant correlations between speech-in-noise recognition and (a) behavioral AMFD thresholds and (b) AMD-ACC peak amplitudes. AMD and AMFD hold potential to develop a clinically applicable tool for assessing TEM encoding to predict speech recognition in CI users.


Subject(s)
Cochlear Implantation , Cochlear Implants , Speech Perception , Adult , Humans , Psychoacoustics , Speech Perception/physiology , Speech , Acoustic Stimulation , Evoked Potentials, Auditory/physiology
8.
Hear Res ; 445: 108992, 2024 Apr.
Article in English | MEDLINE | ID: mdl-38492447

ABSTRACT

Noise sensitivity and hyperacusis are decreased sound tolerance conditions that are not well delineated or defined. This paper presents the correlations and distributions of the Noise Sensitivity Scale (NSS) and the Hyperacusis Questionnaire (HQ) scores in two distinct large samples. In Study 1, a community-based sample of young healthy adults (n = 103) exhibited a strong correlation (r = 0.74) between the two questionnaires. The mean NSS and HQ scores were 54.4 ± 16.9 and 12.5 ± 7.5, respectively. NSS scores displayed a normal distribution, whereas HQ scores showed a slight positive skew. In Study 2, a clinical sample of Veterans with or without clinical comorbidities (n = 95) showed a moderate correlation (r = 0.58) between the two questionnaires. The mean scores were 66.6 ± 15.6 and 15.3 ± 7.3 on the NSS and HQ, respectively. Both questionnaires' scores followed a normal distribution. In both samples, participants who self-identified as having decreased sound tolerance scored higher on both questionnaires. These findings provide reference data from two diverse sample groups. The moderate to strong correlations observed in both studies suggest a significant overlap between noise sensitivity and hyperacusis. The results underscore that NSS and HQ should not be used interchangeably, as they aim to measure distinct constructs, however to what extent they actually do remains to be determined. Further investigation should distinguish between these conditions through a comprehensive psychometric analysis of the questionnaires and a thorough exploration of psychoacoustic, neurological, and physiological differences that set them apart.


Subject(s)
Hyperacusis , Tinnitus , Adult , Humans , Hyperacusis/diagnosis , Surveys and Questionnaires , Sound , Psychoacoustics
9.
J Acoust Soc Am ; 155(2): 901-914, 2024 02 01.
Article in English | MEDLINE | ID: mdl-38310608

ABSTRACT

Dealing with newborns' health is a delicate matter since they cannot express needs, and crying does not reflect their condition. Although newborn cries have been studied for various purposes, there is no prior research on distinguishing a certain pathology from other pathologies so far. Here, an unsophisticated framework is proposed for the study of septic newborns amid a collective of other pathologies. The cry was analyzed with music inspired and speech processing inspired features. Furthermore, neighborhood component analysis (NCA) feature selection was employed with two goals: (i) Exploring how the elements of each feature set contributed to classification outcome; (ii) investigating to what extent the feature space could be compacted. The attained results showed success of both experiments introduced in this study, with 88.66% for the decision template fusion (DTF) technique and a consistent enhancement in comparison to all feature sets in terms of accuracy and 86.22% for the NCA feature selection method by drastically downsizing the feature space from 86 elements to only 6 elements. The achieved results showed great potential for identifying a certain pathology from other pathologies that may have similar effects on the cry patterns as well as proving the success of the proposed framework.


Subject(s)
Crying , Infant, Newborn , Humans , Psychoacoustics
10.
Nat Commun ; 15(1): 1482, 2024 Feb 19.
Article in English | MEDLINE | ID: mdl-38369535

ABSTRACT

The phenomenon of musical consonance is an essential feature in diverse musical styles. The traditional belief, supported by centuries of Western music theory and psychological studies, is that consonance derives from simple (harmonic) frequency ratios between tones and is insensitive to timbre. Here we show through five large-scale behavioral studies, comprising 235,440 human judgments from US and South Korean populations, that harmonic consonance preferences can be reshaped by timbral manipulations, even as far as to induce preferences for inharmonic intervals. We show how such effects may suggest perceptual origins for diverse scale systems ranging from the gamelan's slendro scale to the tuning of Western mean-tone and equal-tempered scales. Through computational modeling we show that these timbral manipulations dissociate competing psychoacoustic mechanisms underlying consonance, and we derive an updated computational model combining liking of harmonicity, disliking of fast beats (roughness), and liking of slow beats. Altogether, this work showcases how large-scale behavioral experiments can inform classical questions in auditory perception.


Subject(s)
Music , Humans , Psychoacoustics , Music/psychology , Auditory Perception , Emotions , Judgment , Acoustic Stimulation
11.
Atten Percept Psychophys ; 86(3): 931-941, 2024 Apr.
Article in English | MEDLINE | ID: mdl-38418807

ABSTRACT

There is an increasing body of evidence suggesting that there are low-level perceptual processes involved in crossmodal correspondences. In this study, we investigate the involvement of the superior colliculi in three basic crossmodal correspondences: elevation/pitch, lightness/pitch, and size/pitch. Using a psychophysical design, we modulate visual input to the superior colliculus to test whether the superior colliculus is required for behavioural crossmodal congruency effects to manifest in an unspeeded multisensory discrimination task. In the elevation/pitch task, superior colliculus involvement is required for a behavioural elevation/pitch congruency effect to manifest in the task. In the lightness/pitch and size/pitch task, we observed a behavioural elevation/pitch congruency effect regardless of superior colliculus involvement. These results suggest that the elevation/pitch correspondence may be processed differently to other low-level crossmodal correspondences. The implications of a distributed model of crossmodal correspondence processing in the brain are discussed.


Subject(s)
Pattern Recognition, Visual , Superior Colliculi , Humans , Superior Colliculi/physiology , Male , Female , Adult , Young Adult , Pattern Recognition, Visual/physiology , Size Perception/physiology , Attention/physiology , Pitch Discrimination/physiology , Association , Psychoacoustics , Orientation/physiology
12.
J Acoust Soc Am ; 155(1): 68-77, 2024 01 01.
Article in English | MEDLINE | ID: mdl-38174963

ABSTRACT

Hearing-impaired (HI) listeners have been shown to exhibit increased fusion of dichotic vowels, even with different fundamental frequency (F0), leading to binaural spectral averaging and interference. To determine if similar fusion and averaging occurs for consonants, four natural and synthesized stop consonants (/pa/, /ba/, /ka/, /ga/) at three F0s of 74, 106, and 185 Hz were presented dichotically-with ΔF0 varied-to normal-hearing (NH) and HI listeners. Listeners identified the one or two consonants perceived, and response options included /ta/ and /da/ as fused percepts. As ΔF0 increased, both groups showed decreases in fusion and increases in percent correct identification of both consonants, with HI listeners displaying similar fusion but poorer identification. Both groups exhibited spectral averaging (psychoacoustic fusion) of place of articulation but phonetic feature fusion for differences in voicing. With synthetic consonants, NH subjects showed increased fusion and decreased identification. Most HI listeners were unable to discriminate the synthetic consonants. The findings suggest smaller differences between groups in consonant fusion than vowel fusion, possibly due to the presence of more cues for segregation in natural speech or reduced reliance on spectral cues for consonant perception. The inability of HI listeners to discriminate synthetic consonants suggests a reliance on cues other than formant transitions for consonant discrimination.


Subject(s)
Hearing Loss, Sensorineural , Hearing Loss , Speech Perception , Humans , Speech Perception/physiology , Psychoacoustics , Phonetics , Hearing
13.
Int J Audiol ; 63(2): 127-135, 2024 Feb.
Article in English | MEDLINE | ID: mdl-36633444

ABSTRACT

OBJECTIVE: The aim of this study was to investigate whether consumer-grade mobile audio equipment can be reliably used as a platform for the notched-noise test, including when the test is conducted outside the laboratory. DESIGN: Two studies were conducted: Study 1 was a notched-noise masking experiment with three different setups: in a psychoacoustic test booth with a standard laboratory PC; in a psychoacoustic test booth with a mobile device; and in a quiet office room with a mobile device. Study 2 employed the same task as Study 1, but compared circumaural headphones to insert earphones. STUDY SAMPLE: Nine and ten young, normal-hearing participants completed studies 1 and 2, respectively. RESULTS: The test-retest accuracy of the notched-noise test on the mobile implementation did not differ from that for the laboratory setup. A possible effect of the earphone design was identified in Study 1, which was corroborated by Study 2, where test-retest variability was smallest when comparing results from experiments conducted using identical acoustic transducers. CONCLUSIONS: Results and test-retest repeatability comparable to standard laboratory settings for the notched-noise test can be obtained with mobile equipment outside the laboratory.


Subject(s)
Acoustics , Noise , Humans , Noise/adverse effects , Psychoacoustics , Computers, Handheld , Transducers
14.
Behav Res Methods ; 56(3): 1433-1448, 2024 Mar.
Article in English | MEDLINE | ID: mdl-37326771

ABSTRACT

Anonymous web-based experiments are increasingly used in many domains of behavioral research. However, online studies of auditory perception, especially of psychoacoustic phenomena pertaining to low-level sensory processing, are challenging because of limited available control of the acoustics, and the inability to perform audiometry to confirm normal-hearing status of participants. Here, we outline our approach to mitigate these challenges and validate our procedures by comparing web-based measurements to lab-based data on a range of classic psychoacoustic tasks. Individual tasks were created using jsPsych, an open-source JavaScript front-end library. Dynamic sequences of psychoacoustic tasks were implemented using Django, an open-source library for web applications, and combined with consent pages, questionnaires, and debriefing pages. Subjects were recruited via Prolific, a subject recruitment platform for web-based studies. Guided by a meta-analysis of lab-based data, we developed and validated a screening procedure to select participants for (putative) normal-hearing status based on their responses in a suprathreshold task and a survey. Headphone use was standardized by supplementing procedures from prior literature with a binaural hearing task. Individuals meeting all criteria were re-invited to complete a range of classic psychoacoustic tasks. For the re-invited participants, absolute thresholds were in excellent agreement with lab-based data for fundamental frequency discrimination, gap detection, and sensitivity to interaural time delay and level difference. Furthermore, word identification scores, consonant confusion patterns, and co-modulation masking release effect also matched lab-based studies. Our results suggest that web-based psychoacoustics is a viable complement to lab-based research. Source code for our infrastructure is provided.


Subject(s)
Auditory Perception , Hearing , Humans , Psychoacoustics , Hearing/physiology , Auditory Perception/physiology , Audiometry , Internet , Auditory Threshold/physiology , Acoustic Stimulation
15.
Augment Altern Commun ; 40(1): 57-67, 2024 Mar.
Article in English | MEDLINE | ID: mdl-37811949

ABSTRACT

Children with complex communication needs often have multiple disabilities including visual impairments that impact their ability to interact with aided augmentative and alternative communication (AAC) systems. Just as the field benefited from a consideration of visual cognitive neuroscience in construction of visual displays, an exploration of psychoacoustics can potentially assist in maximizing the possibilities within AAC systems when the visual channel is either (a) not the primary sensory mode, or (b) is one that can be augmented to ultimately benefit AAC outcomes. The purpose of this paper is to highlight background information about psychoacoustics and present possible future directions for the design of aided AAC system technologies for children with visual impairments who rely on auditory information to learn and utilize AAC.


Subject(s)
Communication Aids for Disabled , Communication Disorders , Child , Humans , Psychoacoustics , Communication Disorders/psychology , Communication , Vision Disorders
16.
PLoS One ; 18(12): e0294645, 2023.
Article in English | MEDLINE | ID: mdl-38051728

ABSTRACT

There is debate whether the foundations of consonance and dissonance are rooted in culture or in psychoacoustics. In order to disentangle the contribution of culture and psychoacoustics, we considered automatic responses to the perfect fifth and the major second (flattened by 25 cents) intervals alongside conscious evaluations of the same intervals across two cultures and two levels of musical expertise. Four groups of participants completed the tasks: expert performers of Lithuanian Sutartines, English speaking musicians in Western diatonic genres, Lithuanian non-musicians and English-speaking non-musicians. Sutartines singers were chosen as this style of singing is an example of 'beat diaphony' where intervals of parts form predominantly rough sonorities and audible beats. There was no difference in automatic responses to intervals, suggesting that an aversion to acoustically rough intervals is not governed by cultural familiarity but may have a physical basis in how the human auditory system works. However, conscious evaluations resulted in group differences with Sutartines singers rating both the flattened major as more positive than did other groups. The results are discussed in the context of recent developments in consonance and dissonance research.


Subject(s)
Music , Singing , Humans , Psychoacoustics , Recognition, Psychology , Consciousness , Acoustic Stimulation/methods , Auditory Perception/physiology
17.
J Acoust Soc Am ; 154(4): 1963-1976, 2023 10 01.
Article in English | MEDLINE | ID: mdl-37782118

ABSTRACT

Appropriate sound quality models for noise-induced discomfort are necessary for a better acoustic comfort design in the aircraft cabin. This study investigates the acoustic discomfort in two large passenger aeroplanes (i.e., planes A and B). We recorded the noise at 21 positions in each aircraft cabin and selected 42 stimuli ranging from 72 to 81 dB(A) during the cruising flights. Twenty-four participants rated the noise discomfort by the absolute magnitude estimation method. The discomfort values in the middle section of the aircraft cabin are 10% points higher than in the front or rear section. The discomfort magnitude was dominated by loudness and influenced by roughness and sharpness. A multiple linear (MA) discomfort model was established, accounting for the relationship between the discomfort and sound quality metrics (i.e., loudness, sharpness, and roughness). The MA model estimated noise discomfort better than the Zwicker and other (i.e., More and Di) psychoacoustic annoyance (PA) models. We modified the coefficients of independent variables in the formulations of Zwicker, Di, and More PA models, respectively, according to the present experimental results. The correlation coefficients between the estimated and measured values of the modified models were at least 20% points higher than the original ones.


Subject(s)
Noise , Sound , Humans , Psychoacoustics , Linear Models , Aircraft
18.
J Acoust Soc Am ; 154(3): 1770-1786, 2023 09 01.
Article in English | MEDLINE | ID: mdl-37721403

ABSTRACT

Both dynamic variation of interaural time difference (ITD) and static spectral cues provide information for front-back discrimination and vertical localization. However, the contributions of the two cues are still unclear. The static spectral cue has conventionally been regarded as the dominant one. In the present work, psychoacoustic experiments were conducted to examine the contribution of dynamic ITD and static spectral cues to vertical localization in the median plane. By modifying the head-related transfer functions used in a dynamic virtual auditory display, binaural signals with conflicting dynamic ITD and spectral cues that were either static or dynamically modified according to instantaneous head position were created. The results indicated that the dynamic ITD and static spectral cues contribute to vertical localization at low and high frequencies, respectively. For full a bandwidth stimulus, conflicting dynamic ITD and static spectral cues usually result in two separated virtual sources at different elevations corresponding to the spatial information conveyed by the low- and high-frequency bands, respectively. In most cases, no fused localization occurs in the high-level cognition system. Therefore, dynamic ITD and static spectral cues contribute to vertical localization at different frequency ranges, and neither of them dominates vertical localization in the case of wideband stimuli.


Subject(s)
Cues , Sound Localization , Cognition , Psychoacoustics
19.
Eur J Neurosci ; 58(7): 3686-3704, 2023 10.
Article in English | MEDLINE | ID: mdl-37752605

ABSTRACT

Human listeners prefer octave intervals slightly above the exact 2:1 frequency ratio. To study the neural underpinnings of this subjective preference, called the octave enlargement phenomenon, we compared neural responses between exact, slightly enlarged, oversized, and compressed octaves (or their multiples). The first experiment (n = 20) focused on the N1 and P2 event-related potentials (ERPs) elicited in EEG 50-250 ms after the second tone onset during passive listening of one-octave intervals. In the second experiment (n = 20) applying four-octave intervals, musician participants actively rated the different octave types as 'low', 'good' and 'high'. The preferred slightly enlarged octave was individually determined prior to the second experiment. In both experiments, N1-P2 peak-to-peak amplitudes attenuated for the exact and slightly enlarged octave intervals compared with compressed and oversized intervals, suggesting overlapping neural representations of tones an octave (or its multiples) apart. While there were no differences between the N1-P2 amplitudes to the exact and preferred enlarged octaves, ERP amplitudes differed after 500 ms from onset of the second tone of the pair. In the multivariate pattern analysis (MVPA) of the second experiment, the different octave types were distinguishable (spatial classification across electroencephalography [EEG] channels) 200 ms after second tone onset. Temporal classification within channels suggested two separate discrimination processes peaking around 300 and 700 ms. These findings appear to be related to active listening, as no multivariate results were found in the first, passive listening experiment. The present results suggest that the subjectively preferred octave size is resolved at the late stages of auditory processing.


Subject(s)
Evoked Potentials , Music , Humans , Psychoacoustics , Electroencephalography , Auditory Perception/physiology , Evoked Potentials, Auditory/physiology , Acoustic Stimulation
20.
J Acoust Soc Am ; 154(2): 751-762, 2023 08 01.
Article in English | MEDLINE | ID: mdl-37556566

ABSTRACT

Web-based testing is an appealing option for expanding psychoacoustics research outside laboratory environments due to its simple logistics. For example, research participants partake in listening tasks using their own computer and audio hardware and can participate in a comfortable environment of their choice at their own pace. However, it is unknown how deviations from conventional in-lab testing affect data quality, particularly in binaural hearing tasks that traditionally require highly precise audio presentation. Here, we used an online platform to replicate two published in-lab experiments: lateralization to interaural time and level differences (ITD and ILD, experiment I) and dichotic and contralateral unmasking of speech (experiment II) in normal-hearing (NH) young adults. Lateralization data collected online were strikingly similar to in-lab results. Likewise, the amount of unmasking measured online and in-lab differed by less than 1 dB, although online participants demonstrated higher speech reception thresholds overall than those tested in-lab by up to ∼7 dB. Results from online participants who completed a hearing screening versus those who self-reported NH did not differ significantly. We conclude that web-based psychoacoustics testing is a viable option for assessing binaural hearing abilities among young NH adults and discuss important considerations for online study design.


Subject(s)
Speech Perception , Young Adult , Humans , Psychoacoustics , Hearing , Auditory Perception , Internet
SELECTION OF CITATIONS
SEARCH DETAIL