Your browser doesn't support javascript.
loading
: 20 | 50 | 100
1 - 20 de 79
1.
Front Robot AI ; 11: 1393637, 2024.
Article En | MEDLINE | ID: mdl-38835930

We recently developed a biomimetic robotic eye with six independent tendons, each controlled by their own rotatory motor, and with insertions on the eye ball that faithfully mimic the biomechanics of the human eye. We constructed an accurate physical computational model of this system, and learned to control its nonlinear dynamics by optimising a cost that penalised saccade inaccuracy, movement duration, and total energy expenditure of the motors. To speed up the calculations, the physical simulator was approximated by a recurrent neural network (NARX). We showed that the system can produce realistic eye movements that closely resemble human saccades in all directions: their nonlinear main-sequence dynamics (amplitude-peak eye velocity and duration relationships), cross-coupling of the horizontal and vertical movement components leading to approximately straight saccade trajectories, and the 3D kinematics that restrict 3D eye orientations to a plane (Listing's law). Interestingly, the control algorithm had organised the motors into appropriate agonist-antagonist muscle pairs, and the motor signals for the eye resembled the well-known pulse-step characteristics that have been reported for monkey motoneuronal activity. We here fully analyse the eye-movement properties produced by the computational model across the entire oculomotor range and the underlying control signals. We argue that our system may shed new light on the neural control signals and their couplings within the final neural pathways of the primate oculomotor system, and that an optimal control principle may account for a wide variety of oculomotor behaviours. The generated data are publicly available at https://data.ru.nl/collections/di/dcn/DSC_626870_0003_600.

2.
Invest Ophthalmol Vis Sci ; 65(5): 39, 2024 May 01.
Article En | MEDLINE | ID: mdl-38787546

Purpose: Post-saccadic oscillations (PSOs) reflect movements of gaze that result from motion of the pupil and lens relative to the eyeball rather than eyeball rotations. Here, we analyzed the characteristics of PSOs in subjects with age-related macular degeneration (AMD), retinitis pigmentosa (RP), and normal vision (NV). Our aim was to assess the differences in PSOs between people with vision loss and healthy controls because PSOs affect retinal image stability after each saccade. Methods: Participants completed a horizontal saccade task and their gaze was measured using a pupil-based eye tracker. Oscillations occurring in the 80 to 200 ms post-saccadic period were described with a damped oscillation model. We compared the amplitude, decay time constant, and frequency of the PSOs for the three different groups. We also examined the correlation between these PSO parameters and the amplitude, peak velocity, and final deceleration of the preceding saccades. Results: Subjects with vision loss (AMD, n = 6, and RP, n = 5) had larger oscillation amplitudes, longer decay constants, and lower frequencies than subjects with NV (n = 7). The oscillation amplitudes increased with increases in saccade deceleration in all three groups. The other PSO parameters, however, did not show consistent correlations with either saccade amplitude or peak velocity. Conclusions: Post-saccadic fixation stability in AMD and RP is reduced due to abnormal PSOs. The differences with respect to NV are not due to differences in saccade kinematics, suggesting that anatomic and neuronal variations affect the suspension of the iris and the lens in the patients' eyes.


Fixation, Ocular , Macular Degeneration , Pupil , Retinitis Pigmentosa , Saccades , Humans , Saccades/physiology , Retinitis Pigmentosa/physiopathology , Female , Male , Fixation, Ocular/physiology , Middle Aged , Macular Degeneration/physiopathology , Aged , Pupil/physiology , Lens, Crystalline/physiopathology , Adult , Visual Acuity/physiology
3.
J Neurophysiol ; 131(1): 38-63, 2024 Jan 01.
Article En | MEDLINE | ID: mdl-37965933

Human speech and vocalizations in animals are rich in joint spectrotemporal (S-T) modulations, wherein acoustic changes in both frequency and time are functionally related. In principle, the primate auditory system could process these complex dynamic sounds based on either an inseparable representation of S-T features or, alternatively, a separable representation. The separability hypothesis implies an independent processing of spectral and temporal modulations. We collected comparative data on the S-T hearing sensitivity in humans and macaque monkeys to a wide range of broadband dynamic spectrotemporal ripple stimuli employing a yes-no signal-detection task. Ripples were systematically varied, as a function of density (spectral modulation frequency), velocity (temporal modulation frequency), or modulation depth, to cover a listener's full S-T modulation sensitivity, derived from a total of 87 psychometric ripple detection curves. Audiograms were measured to control for normal hearing. Determined were hearing thresholds, reaction time distributions, and S-T modulation transfer functions (MTFs), both at the ripple detection thresholds and at suprathreshold modulation depths. Our psychophysically derived MTFs are consistent with the hypothesis that both monkeys and humans employ analogous perceptual strategies: S-T acoustic information is primarily processed separable. Singular value decomposition (SVD), however, revealed a small, but consistent, inseparable spectral-temporal interaction. Finally, SVD analysis of the known visual spatiotemporal contrast sensitivity function (CSF) highlights that human vision is space-time inseparable to a much larger extent than is the case for S-T sensitivity in hearing. Thus, the specificity with which the primate brain encodes natural sounds appears to be less strict than is required to adequately deal with natural images.NEW & NOTEWORTHY We provide comparative data on primate audition of naturalistic sounds comprising hearing thresholds, reaction time distributions, and spectral-temporal modulation transfer functions. Our psychophysical experiments demonstrate that auditory information is primarily processed in a spectral-temporal-independent manner by both monkeys and humans. Singular value decomposition of known visual spatiotemporal contrast sensitivity, in comparison to our auditory spectral-temporal sensitivity, revealed a striking contrast in how the brain encodes natural sounds as opposed to natural images, as vision appears to be space-time inseparable.


Speech Perception , Time Perception , Animals , Humans , Haplorhini , Auditory Perception , Hearing , Acoustic Stimulation/methods
4.
Commun Biol ; 6(1): 927, 2023 09 09.
Article En | MEDLINE | ID: mdl-37689726

The midbrain superior colliculus is a crucial sensorimotor stage for programming and generating saccadic eye-head gaze shifts. Although it is well established that superior colliculus cells encode a neural command that specifies the amplitude and direction of the upcoming gaze-shift vector, there is controversy about the role of the firing-rate dynamics of these neurons during saccades. In our earlier work, we proposed a simple quantitative model that explains how the recruited superior colliculus population may specify the detailed kinematics (trajectories and velocity profiles) of head-restrained saccadic eye movements. We here show that the same principles may apply to a wide range of saccadic eye-head gaze shifts with strongly varying kinematics, despite the substantial nonlinearities and redundancy in programming and execute rapid goal-directed eye-head gaze shifts to peripheral targets. Our findings could provide additional evidence for an important role of the superior colliculus in the optimal control of saccades.


Neurons , Superior Colliculi , Biomechanical Phenomena , Fixation, Ocular , Saccades
5.
Front Neurosci ; 17: 1183126, 2023.
Article En | MEDLINE | ID: mdl-37521701

A cochlear implant (CI) is a neurotechnological device that restores total sensorineural hearing loss. It contains a sophisticated speech processor that analyzes and transforms the acoustic input. It distributes its time-enveloped spectral content to the auditory nerve as electrical pulsed stimulation trains of selected frequency channels on a multi-contact electrode that is surgically inserted in the cochlear duct. This remarkable brain interface enables the deaf to regain hearing and understand speech. However, tuning of the large (>50) number of parameters of the speech processor, so-called "device fitting," is a tedious and complex process, which is mainly carried out in the clinic through 'one-size-fits-all' procedures. Current fitting typically relies on limited and often subjective data that must be collected in limited time. Despite the success of the CI as a hearing-restoration device, variability in speech-recognition scores among users is still very large, and mostly unexplained. The major factors that underly this variability incorporate three levels: (i) variability in auditory-system malfunction of CI-users, (ii) variability in the selectivity of electrode-to-auditory nerve (EL-AN) activation, and (iii) lack of objective perceptual measures to optimize the fitting. We argue that variability in speech recognition can only be alleviated by using objective patient-specific data for an individualized fitting procedure, which incorporates knowledge from all three levels. In this paper, we propose a series of experiments, aimed at collecting a large amount of objective (i.e., quantitative, reproducible, and reliable) data that characterize the three processing levels of the user's auditory system. Machine-learning algorithms that process these data will eventually enable the clinician to derive reliable and personalized characteristics of the user's auditory system, the quality of EL-AN signal transfer, and predictions of the perceptual effects of changes in the current fitting.

6.
Trends Hear ; 27: 23312165221143907, 2023.
Article En | MEDLINE | ID: mdl-36605011

Many cochlear implant users with binaural residual (acoustic) hearing benefit from combining electric and acoustic stimulation (EAS) in the implanted ear with acoustic amplification in the other. These bimodal EAS listeners can potentially use low-frequency binaural cues to localize sounds. However, their hearing is generally asymmetric for mid- and high-frequency sounds, perturbing or even abolishing binaural cues. Here, we investigated the effect of a frequency-dependent binaural asymmetry in hearing thresholds on sound localization by seven bimodal EAS listeners. Frequency dependence was probed by presenting sounds with power in low-, mid-, high-, or mid-to-high-frequency bands. Frequency-dependent hearing asymmetry was present in the bimodal EAS listening condition (when using both devices) but was also induced by independently switching devices on or off. Using both devices, hearing was near symmetric for low frequencies, asymmetric for mid frequencies with better hearing thresholds in the implanted ear, and monaural for high frequencies with no hearing in the non-implanted ear. Results show that sound-localization performance was poor in general. Typically, localization was strongly biased toward the better hearing ear. We observed that hearing asymmetry was a good predictor for these biases. Notably, even when hearing was symmetric a preferential bias toward the ear using the hearing aid was revealed. We discuss how frequency dependence of any hearing asymmetry may lead to binaural cues that are spatially inconsistent as the spectrum of a sound changes. We speculate that this inconsistency may prevent accurate sound-localization even after long-term exposure to the hearing asymmetry.


Cochlear Implantation , Cochlear Implants , Hearing Aids , Sound Localization , Speech Perception , Humans , Speech Perception/physiology , Cochlear Implantation/methods , Hearing , Sound Localization/physiology , Acoustic Stimulation/methods
7.
Front Comput Neurosci ; 16: 1040646, 2022.
Article En | MEDLINE | ID: mdl-36465967

Introduction: To reorient gaze (the eye's direction in space) towards a target is an overdetermined problem, as infinitely many combinations of eye- and head movements can specify the same gaze-displacement vector. Yet, behavioral measurements show that the primate gaze-control system selects a specific contribution of eye- and head movements to the saccade, which depends on the initial eye-in-head orientation. Single-unit recordings in the primate superior colliculus (SC) during head-unrestrained gaze shifts have further suggested that cells may encode the instantaneous trajectory of a desired straight gaze path in a feedforward way by the total cumulative number of spikes in the neural population, and that the instantaneous gaze kinematics are thus determined by the neural firing rates. The recordings also indicated that the latter is modulated by the initial eye position. We recently proposed a conceptual model that accounts for many of the observed properties of eye-head gaze shifts and on the potential role of the SC in gaze control. Methods: Here, we extend and test the model by incorporating a spiking neural network of the SC motor map, the output of which drives the eye-head motor control circuitry by linear cumulative summation of individual spike effects of each recruited SC neuron. We propose a simple neural mechanism on SC cells that explains the modulatory influence of feedback from an initial eye-in-head position signal on their spiking activity. The same signal also determines the onset delay of the head movement with respect to the eye. Moreover, the downstream eye- and head burst generators were taken to be linear, as our earlier work had indicated that much of the non-linear main-sequence kinematics of saccadic eye movements may be due to neural encoding at the collicular level, rather than at the brainstem. Results and discussion: We investigate how the spiking activity of the SC population drives gaze to the intended target location within a dynamic local gaze-velocity feedback circuit that yields realistic eye- and head-movement kinematics and dynamic SC gaze-movement fields.

8.
Trends Hear ; 26: 23312165221127589, 2022.
Article En | MEDLINE | ID: mdl-36172759

We tested whether sensitivity to acoustic spectrotemporal modulations can be observed from reaction times for normal-hearing and impaired-hearing conditions. In a manual reaction-time task, normal-hearing listeners had to detect the onset of a ripple (with density between 0-8 cycles/octave and a fixed modulation depth of 50%), that moved up or down the log-frequency axis at constant velocity (between 0-64 Hz), in an otherwise-unmodulated broadband white-noise. Spectral and temporal modulations elicited band-pass filtered sensitivity characteristics, with fastest detection rates around 1 cycle/oct and 32 Hz for normal-hearing conditions. These results closely resemble data from other studies that typically used the modulation-depth threshold as a sensitivity criterion. To simulate hearing-impairment, stimuli were processed with a 6-channel cochlear-implant vocoder, and a hearing-aid simulation that introduced separate spectral smearing and low-pass filtering. Reaction times were always much slower compared to normal hearing, especially for the highest spectral densities. Binaural performance was predicted well by the benchmark race model of binaural independence, which models statistical facilitation of independent monaural channels. For the impaired-hearing simulations this implied a "best-of-both-worlds" principle in which the listeners relied on the hearing-aid ear to detect spectral modulations, and on the cochlear-implant ear for temporal-modulation detection. Although singular-value decomposition indicated that the joint spectrotemporal sensitivity matrix could be largely reconstructed from independent temporal and spectral sensitivity functions, in line with time-spectrum separability, a substantial inseparable spectral-temporal interaction was present in all hearing conditions. These results suggest that the reaction-time task yields a valid and effective objective measure of acoustic spectrotemporal-modulation sensitivity.


Cochlear Implantation , Cochlear Implants , Speech Perception , Acoustic Stimulation , Auditory Threshold , Hearing , Humans , Reaction Time
9.
Sci Rep ; 12(1): 6916, 2022 04 28.
Article En | MEDLINE | ID: mdl-35484389

Previous studies have indicated that the location of a large neural population in the Superior Colliculus (SC) motor map specifies the amplitude and direction of the saccadic eye-movement vector, while the saccade trajectory and velocity profile are encoded by the population firing rates. We recently proposed a simple spiking neural network model of the SC motor map, based on linear summation of individual spike effects of each recruited neuron, which accounts for many of the observed properties of SC cells in relation to the ensuing eye movement. However, in the model, the cortical input was kept invariant across different saccades. Electrical microstimulation and reversible lesion studies have demonstrated that the saccade properties are quite robust against large changes in supra-threshold SC activation, but that saccade amplitude and peak eye-velocity systematically decrease at low input strengths. These features were not accounted for by the linear spike-vector summation model. Here we show that the model's input projection strengths and intra-collicular lateral connections can be tuned to generate saccades and neural spiking patterns that closely follow the experimental results.


Saccades , Superior Colliculi , Eye Movements , Neural Networks, Computer , Neurons/physiology , Superior Colliculi/physiology
10.
Sci Rep ; 12(1): 5262, 2022 03 28.
Article En | MEDLINE | ID: mdl-35347172

Several studies have proposed that an optimal speed-accuracy tradeoff underlies the stereotyped relationship between amplitude, duration and peak velocity of saccades (main sequence). To test this theory, we asked 8 participants to make saccades to Gaussian-blurred spots and manipulated the task's accuracy constraints by varying target size (1, 3, and 5°). The largest targets indeed yielded more endpoint scatter (and lower gains) than the smallest targets, although this effect subsided with target eccentricity. The main sequence depended on several interacting factors: saccade latency, saccade gain and target size. Early saccades, which were faster than amplitude-matched late saccades, followed the target-size dependency one would expect from a speed-accuracy tradeoff process. They had higher peak velocities and shorter durations for larger targets than for smaller targets. For late saccades, however, the opposite was found. Deviations from the main sequence also covaried with saccade gain, in line with the idea that motor noise underlies part of the endpoint variability. Thus, our data provide partial evidence that the saccadic system weighs the detrimental effects of motor noise on saccade accuracy against movement duration and speed, but other factors also modulate the kinematics. We discuss the possible involvement of parallel saccade pathways to account for our findings.


Noise , Saccades , Biomechanical Phenomena , Data Collection , Humans , Time Factors
11.
J Speech Lang Hear Res ; 64(12): 5000-5013, 2021 12 13.
Article En | MEDLINE | ID: mdl-34714704

PURPOSE: Speech understanding in noise and horizontal sound localization is poor in most cochlear implant (CI) users with a hearing aid (bimodal stimulation). This study investigated the effect of static and less-extreme adaptive frequency compression in hearing aids on spatial hearing. By means of frequency compression, we aimed to restore high-frequency audibility, and thus improve sound localization and spatial speech recognition. METHOD: Sound-detection thresholds, sound localization, and spatial speech recognition were measured in eight bimodal CI users, with and without frequency compression. We tested two compression algorithms: a static algorithm, which compressed frequencies beyond the compression knee point (160 or 480 Hz), and an adaptive algorithm, which aimed to compress only consonants leaving vowels unaffected (adaptive knee-point frequencies from 736 to 2946 Hz). RESULTS: Compression yielded a strong audibility benefit (high-frequency thresholds improved by 40 and 24 dB for static and adaptive compression, respectively), no meaningful improvement in localization performance (errors remained > 30 deg), and spatial speech recognition across all participants. Localization biases without compression (toward the hearing-aid and implant side for low- and high-frequency sounds, respectively) disappeared or reversed with compression. The audibility benefits provided to each bimodal user partially explained any individual improvements in localization performance; shifts in bias; and, for six out of eight participants, benefits in spatial speech recognition. CONCLUSIONS: We speculate that limiting factors such as a persistent hearing asymmetry and mismatch in spectral overlap prevent compression in bimodal users from improving sound localization. Therefore, the benefit in spatial release from masking by compression is likely due to a shift of attention to the ear with the better signal-to-noise ratio facilitated by compression, rather than an improved spatial selectivity. Supplemental Material https://doi.org/10.23641/asha.16869485.


Cochlear Implantation , Cochlear Implants , Hearing Aids , Sound Localization , Speech Perception , Hearing , Humans , Speech Perception/physiology
12.
Front Neurosci ; 15: 683804, 2021.
Article En | MEDLINE | ID: mdl-34393707

The cochlear implant (CI) allows profoundly deaf individuals to partially recover hearing. Still, due to the coarse acoustic information provided by the implant, CI users have considerable difficulties in recognizing speech, especially in noisy environments. CI users therefore rely heavily on visual cues to augment speech recognition, more so than normal-hearing individuals. However, it is unknown how attention to one (focused) or both (divided) modalities plays a role in multisensory speech recognition. Here we show that unisensory speech listening and reading were negatively impacted in divided-attention tasks for CI users-but not for normal-hearing individuals. Our psychophysical experiments revealed that, as expected, listening thresholds were consistently better for the normal-hearing, while lipreading thresholds were largely similar for the two groups. Moreover, audiovisual speech recognition for normal-hearing individuals could be described well by probabilistic summation of auditory and visual speech recognition, while CI users were better integrators than expected from statistical facilitation alone. Our results suggest that this benefit in integration comes at a cost. Unisensory speech recognition is degraded for CI users when attention needs to be divided across modalities. We conjecture that CI users exhibit an integration-attention trade-off. They focus solely on a single modality during focused-attention tasks, but need to divide their limited attentional resources in situations with uncertainty about the upcoming stimulus modality. We argue that in order to determine the benefit of a CI for speech recognition, situational factors need to be discounted by presenting speech in realistic or complex audiovisual environments.

13.
PLoS Comput Biol ; 17(5): e1008975, 2021 05.
Article En | MEDLINE | ID: mdl-34029310

An interesting problem for the human saccadic eye-movement system is how to deal with the degrees-of-freedom problem: the six extra-ocular muscles provide three rotational degrees of freedom, while only two are needed to point gaze at any direction. Measurements show that 3D eye orientations during head-fixed saccades in far-viewing conditions lie in Listing's plane (LP), in which the eye's cyclotorsion is zero (Listing's law, LL). Moreover, while saccades are executed as single-axis rotations around a stable eye-angular velocity axis, they follow straight trajectories in LP. Another distinctive saccade property is their nonlinear main-sequence dynamics: the affine relationship between saccade size and movement duration, and the saturation of peak velocity with amplitude. To explain all these properties, we developed a computational model, based on a simplified and upscaled robotic prototype of an eye with 3 degrees of freedom, driven by three independent motor commands, coupled to three antagonistic elastic muscle pairs. As the robotic prototype was not intended to faithfully mimic the detailed biomechanics of the human eye, we did not impose specific prior mechanical constraints on the ocular plant that could, by themselves, generate Listing's law and the main-sequence. Instead, our goal was to study how these properties can emerge from the application of optimal control principles to simplified eye models. We performed a numerical linearization of the nonlinear system dynamics around the origin using system identification techniques, and developed open-loop controllers for 3D saccade generation. Applying optimal control to the simulated model, could reproduce both Listing's law and and the main-sequence. We verified the contribution of different terms in the cost optimization functional to realistic 3D saccade behavior, and identified four essential terms: total energy expenditure by the motors, movement duration, gaze accuracy, and the total static force exerted by the muscles during fixation. Our findings suggest that Listing's law, as well as the saccade dynamics and their trajectories, may all emerge from the same common mechanism that aims to optimize speed-accuracy trade-off for saccades, while minimizing the total muscle force during eccentric fixation.


Models, Biological , Saccades , Biomechanical Phenomena , Humans , Orientation , Vision, Ocular
14.
eNeuro ; 8(3)2021.
Article En | MEDLINE | ID: mdl-33875456

Although moving sound-sources abound in natural auditory scenes, it is not clear how the human brain processes auditory motion. Previous studies have indicated that, although ocular localization responses to stationary sounds are quite accurate, ocular smooth pursuit of moving sounds is very poor. We here demonstrate that human subjects faithfully track a sound's unpredictable movements in the horizontal plane with smooth-pursuit responses of the head. Our analysis revealed that the stimulus-response relation was well described by an under-damped passive, second-order low-pass filter in series with an idiosyncratic, fixed, pure delay. The model contained only two free parameters: the system's damping coefficient, and its central (resonance) frequency. We found that the latter remained constant at ∼0.6 Hz throughout the experiment for all subjects. Interestingly, the damping coefficient systematically increased with trial number, suggesting the presence of an adaptive mechanism in the auditory pursuit system (APS). This mechanism functions even for unpredictable sound-motion trajectories endowed with fixed, but covert, frequency characteristics in open-loop tracking conditions. We conjecture that the APS optimizes a trade-off between response speed and effort. Taken together, our data support the existence of a pursuit system for auditory head-tracking, which would suggest the presence of a neural representation of a spatial auditory fovea (AF).


Motion Perception , Pursuit, Smooth , Humans , Movement , Psychomotor Performance , Reaction Time , Sound
15.
Sci Rep ; 11(1): 2150, 2021 01 25.
Article En | MEDLINE | ID: mdl-33495484

The latency of the auditory steady-state response (ASSR) may provide valuable information regarding the integrity of the auditory system, as it could potentially reveal the presence of multiple intracerebral sources. To estimate multiple latencies from high-order ASSRs, we propose a novel two-stage procedure that consists of a nonparametric estimation method, called apparent latency from phase coherence (ALPC), followed by a heuristic sequential forward selection algorithm (SFS). Compared with existing methods, ALPC-SFS requires few prior assumptions, and is straightforward to implement for higher-order nonlinear responses to multi-cosine sound complexes with their initial phases set to zero. It systematically evaluates the nonlinear components of the ASSRs by estimating multiple latencies, automatically identifies involved ASSR components, and reports a latency consistency index. To verify the proposed method, we performed simulations for several scenarios: two nonlinear subsystems with different or overlapping outputs. We compared the results from our method with predictions from existing, parametric methods. We also recorded the EEG from ten normal-hearing adults by bilaterally presenting superimposed tones with four frequencies that evoke a unique set of ASSRs. From these ASSRs, two major latencies were found to be stable across subjects on repeated measurement days. The two latencies are dominated by low-frequency (LF) (near 40 Hz, at around 41-52 ms) and high-frequency (HF) (> 80 Hz, at around 21-27 ms) ASSR components. The frontal-central brain region showed longer latencies on LF components, but shorter latencies on HF components, when compared with temporal-lobe regions. In conclusion, the proposed nonparametric ALPC-SFS method, applied to zero-phase, multi-cosine sound complexes is more suitable for evaluating embedded nonlinear systems underlying ASSRs than existing methods. It may therefore be a promising objective measure for hearing performance and auditory cortex (dys)function.


Auditory Perception/physiology , Electroencephalography , Acoustic Stimulation , Adult , Algorithms , Humans , Signal Processing, Computer-Assisted , Signal-To-Noise Ratio
16.
J Neurophysiol ; 125(2): 556-567, 2021 02 01.
Article En | MEDLINE | ID: mdl-33378250

To program a goal-directed response in the presence of acoustic reflections, the audio-motor system should suppress the detection of time-delayed sources. We examined the effects of spatial separation and interstimulus delay on the ability of human listeners to localize a pair of broadband sounds in the horizontal plane. Participants indicated how many sounds were heard and where these were perceived by making one or two head-orienting localization responses. Results suggest that perceptual fusion of the two sounds depends on delay and spatial separation. Leading and lagging stimuli in close spatial proximity required longer stimulus delays to be perceptually separated than those further apart. Whenever participants heard one sound, their localization responses for synchronous sounds were oriented to a weighted average of both source locations. For short delays, responses were directed toward the leading stimulus location. Increasing spatial separation enhanced this effect. For longer delays, responses were again directed toward a weighted average. When participants perceived two sounds, the first and the second response were directed to either of the leading and lagging source locations. Perceived locations were interchanged often in their temporal order (in ∼40% of trials). We show that the percept of two sounds occurring requires sufficient spatiotemporal separation, after which localization can be performed with high accuracy. We propose that the percept of temporal order of two concurrent sounds results from a different process than localization and discuss how dynamic lateral excitatory-inhibitory interactions within a spatial sensorimotor map could explain the findings.NEW & NOTEWORTHY Sound localization requires spectral and temporal processing of implicit acoustic cues, and is seriously challenged when multiple sources coincide closely in space and time. We systematically varied spatial-temporal disparities for two sounds and instructed listeners to generate goal-directed head movements. We found that even when the auditory system has accurate representations of both sources, it still has trouble to decide whether the scene contained one or two sounds, and in which order they appeared.


Sound Localization , Spatial Behavior , Adult , Brain/physiology , Cues , Female , Head Movements , Humans , Male
17.
Front Neurol ; 11: 915, 2020.
Article En | MEDLINE | ID: mdl-33101160

Several studies have demonstrated the advantages of the bilateral vs. unilateral cochlear implantation in listeners with bilateral severe to profound hearing loss. However, it remains unclear to what extent bilaterally implanted listeners have access to binaural cues, e.g., accurate processing of interaural timing differences (ITDs) for low-frequency sounds (<1.5 kHz) and interaural level differences (ILDs) for high frequencies (>3 kHz). We tested 25 adult listeners, bilaterally implanted with MED-EL cochlear implant (CI) devices, with and without fine-structure (FS) temporal processing as encoding strategy in the low-frequency channels. In order to assess whether the ability to process binaural cues was affected by fine-structure processing, we performed psychophysical ILD and ITD sensitivity measurements and free-field sound localization experiments. We compared the results of the bilaterally implanted listeners with different numbers of FS channels. All CI listeners demonstrated good sensitivity to ILDs, but relatively poor to ITD cues. Although there was a large variability in performance, some bilateral CI users showed remarkably good localization skills. The FS coding strategy for bilateral CI hearing did not improve fine-structure ITD processing for spatial hearing on a group level. However, some CI listeners were able to exploit weakly informative temporal cues to improve their low-frequency spatial perception.

18.
Hear Res ; 385: 107847, 2020 01.
Article En | MEDLINE | ID: mdl-31786443

Congenital unilateral conductive hearing loss (UCHL) jeopardizes directional hearing and speech perception in noisy conditions. Potentially, children with congenital UCHL can benefit from fitting a hearing device, such as a bone-conduction device (BCD). However, the literature reports limited benefit from fitting a BCD, and often, surprisingly, relatively good sound localization in the unaided condition is reported. In this study, we hypothesized that the limited benefit with a BCD is related to (i) insufficient access to binaural cues and (ii) relying on monaural spectral pinna cues for sound localization in the horizontal plane. Directional hearing was tested in seventeen children with congenital UCHL (age 6-19) using a percutaneous BCD. Additionally, a mold was placed in the pinna of the normal-hearing ear to diminish direction-dependent spectral pinna cues. Relatively good localization in azimuth was found in the unaided hearing condition in the majority of the children. Sound localization improved when listening with a BCD, and no correlation between age of implantation and aided localization performance was found. When the mold was inserted, the unaided and aided localization abilities of most children deteriorated. Interestingly, in the children with poor localization performance in the unaided condition, sound localization improved significantly with the BCD, and was hardly affected by molding the pinna of the normal-hearing ear. These observations indicate that the majority of these children rely on spectral pinna cues to localize sounds, independent of listening with or without their device. In conclusion, an important reason for the limited benefit of BCD fitting in children with congenital UCHL might be ascribed to an effective coping strategy (use of spectral pinna cues) that still plays a dominant role after BCD fitting.


Bone Conduction , Congenital Abnormalities/rehabilitation , Correction of Hearing Impairment/instrumentation , Cues , Ear Auricle/physiopathology , Ear/abnormalities , Hearing Aids , Hearing Loss, Conductive/rehabilitation , Hearing Loss, Unilateral/rehabilitation , Persons With Hearing Impairments/rehabilitation , Sound Localization , Adolescent , Child , Congenital Abnormalities/diagnosis , Congenital Abnormalities/physiopathology , Congenital Abnormalities/psychology , Ear/physiopathology , Female , Hearing Loss, Conductive/diagnosis , Hearing Loss, Conductive/physiopathology , Hearing Loss, Conductive/psychology , Hearing Loss, Unilateral/diagnosis , Hearing Loss, Unilateral/physiopathology , Hearing Loss, Unilateral/psychology , Humans , Male , Persons With Hearing Impairments/psychology , Young Adult
19.
Front Hum Neurosci ; 13: 335, 2019.
Article En | MEDLINE | ID: mdl-31611780

We assessed how synchronous speech listening and lipreading affects speech recognition in acoustic noise. In simple audiovisual perceptual tasks, inverse effectiveness is often observed, which holds that the weaker the unimodal stimuli, or the poorer their signal-to-noise ratio, the stronger the audiovisual benefit. So far, however, inverse effectiveness has not been demonstrated for complex audiovisual speech stimuli. Here we assess whether this multisensory integration effect can also be observed for the recognizability of spoken words. To that end, we presented audiovisual sentences to 18 native-Dutch normal-hearing participants, who had to identify the spoken words from a finite list. Speech-recognition performance was determined for auditory-only, visual-only (lipreading), and auditory-visual conditions. To modulate acoustic task difficulty, we systematically varied the auditory signal-to-noise ratio. In line with a commonly observed multisensory enhancement on speech recognition, audiovisual words were more easily recognized than auditory-only words (recognition thresholds of -15 and -12 dB, respectively). We here show that the difficulty of recognizing a particular word, either acoustically or visually, determines the occurrence of inverse effectiveness in audiovisual word integration. Thus, words that are better heard or recognized through lipreading, benefit less from bimodal presentation. Audiovisual performance at the lowest acoustic signal-to-noise ratios (45%) fell below the visual recognition rates (60%), reflecting an actual deterioration of lipreading in the presence of excessive acoustic noise. This suggests that the brain may adopt a strategy in which attention has to be divided between listening and lipreading.

20.
eNeuro ; 6(6)2019.
Article En | MEDLINE | ID: mdl-31601632

Sound localization in the horizontal plane (azimuth) relies mainly on binaural difference cues in sound level and arrival time. Blocking one ear will perturb these cues, and may strongly affect azimuth performance of the listener. However, single-sided deaf listeners, as well as acutely single-sided plugged normal-hearing subjects, often use a combination of (ambiguous) monaural head-shadow cues, impoverished binaural level-difference cues, and (veridical, but limited) pinna- and head-related spectral cues to estimate source azimuth. To what extent listeners can adjust the relative contributions of these different cues is unknown, as the mechanisms underlying adaptive processes to acute monauralization are still unclear. By providing visual feedback during a brief training session with a high-pass (HP) filtered sound at a fixed sound level, we investigated the ability of listeners to adapt to their erroneous sound-localization percepts. We show that acutely plugged listeners rapidly adjusted the relative contributions of perceived sound level, and the spectral and distorted binaural cues, to improve their localization performance in azimuth also for different sound levels and locations than those experienced during training. Interestingly, our results also show that this acute cue-reweighting led to poorer localization performance in elevation, which was in line with the acoustic-spatial information provided during training. We conclude that the human auditory system rapidly readjusts the weighting of all relevant localization cues, to adequately respond to the demands of the current acoustic environment, even if the adjustments may hamper veridical localization performance in the real world.


Adaptation, Physiological/physiology , Auditory Perception/physiology , Feedback, Sensory/physiology , Sound Localization/physiology , Visual Perception/physiology , Acoustic Stimulation , Adult , Female , Humans , Male , Middle Aged , Young Adult
...