RESUMO
Accelerating text input in augmentative and alternative communication (AAC) is a long-standing area of research with bearings on the quality of life in individuals with profound motor impairments. Recent advances in large language models (LLMs) pose opportunities for re-thinking strategies for enhanced text entry in AAC. In this paper, we present SpeakFaster, consisting of an LLM-powered user interface for text entry in a highly-abbreviated form, saving 57% more motor actions than traditional predictive keyboards in offline simulation. A pilot study on a mobile device with 19 non-AAC participants demonstrated motor savings in line with simulation and relatively small changes in typing speed. Lab and field testing on two eye-gaze AAC users with amyotrophic lateral sclerosis demonstrated text-entry rates 29-60% above baselines, due to significant saving of expensive keystrokes based on LLM predictions. These findings form a foundation for further exploration of LLM-assisted text entry in AAC and other user interfaces.
Assuntos
Esclerose Lateral Amiotrófica , Auxiliares de Comunicação para Pessoas com Deficiência , Fixação Ocular , Humanos , Esclerose Lateral Amiotrófica/fisiopatologia , Feminino , Masculino , Projetos Piloto , Fixação Ocular/fisiologia , Idioma , Adulto , Interface Usuário-Computador , Pessoa de Meia-Idade , ComunicaçãoRESUMO
Microsaccades are known to be associated with a deficit in perceptual detection performance for brief probe flashes presented in their temporal vicinity. However, it is still not clear how such a deficit might depend on the visual environment across which microsaccades are generated. Here, and motivated by studies demonstrating an interaction between visual background image appearance and perceptual suppression strength associated with large saccades, we probed peripheral perceptual detection performance of human subjects while they generated microsaccades over three different visual backgrounds. Subjects fixated near the center of a low spatial frequency grating, a high spatial frequency grating, or a small white fixation spot over an otherwise gray background. When a computer process detected a microsaccade, it presented a brief peripheral probe flash at one of four locations (over a uniform gray background) and at different times. After collecting full psychometric curves, we found that both perceptual detection thresholds and slopes of psychometric curves were impaired for peripheral flashes in the immediate temporal vicinity of microsaccades, and they recovered with later flash times. Importantly, the threshold elevations, but not the psychometric slope reductions, were stronger for the white fixation spot than for either of the two gratings. Thus, like with larger saccades, microsaccadic suppression strength can show a certain degree of image dependence. However, unlike with larger saccades, stronger microsaccadic suppression did not occur with low spatial frequency textures. This observation might reflect the different spatiotemporal retinal transients associated with the small microsaccades in our study versus larger saccades.
Assuntos
Estimulação Luminosa , Movimentos Sacádicos , Humanos , Movimentos Sacádicos/fisiologia , Estimulação Luminosa/métodos , Adulto , Masculino , Feminino , Fóvea Central/fisiologia , Limiar Sensorial/fisiologia , Adulto Jovem , Percepção Visual/fisiologia , Campos Visuais/fisiologia , Fixação Ocular/fisiologia , Psicometria/métodosRESUMO
Every day, millions of viewers worldwide engage with subtitled content, and an increasing number choose to watch without sound. In this mixed-methods study, we examine the impact of sound presence or absence on the viewing experience of both first-language (L1) and second-language (L2) viewers when they watch subtitled videos. We explore this novel phenomenon through comprehension and recall post-tests, self-reported cognitive load, immersion, and enjoyment measures, as well as gaze pattern analysis using eye tracking. We also investigate viewers' motivations for opting for audiovisual content without sound and explore how the absence of sound impacts their viewing experience, using in-depth, semi-structured interviews. Our goal is to ascertain whether these effects are consistent among L2 and L1 speakers from different language varieties. To achieve this, we tested L1-British English, L1-Australian English and L2-English (L1-Polish) language speakers (n = 168) while they watched English-language audiovisual material with English subtitles with and without sound. The findings show that when watching videos without sound, viewers experienced increased cognitive load, along with reduced comprehension, immersion and overall enjoyment. Examination of participants' gaze revealed that the absence of sound significantly affected the viewing experience, increasing the need for subtitles and thus increasing the viewers' propensity to process them more thoroughly. The absence of sound emerged as a global constraint that made reading more effortful. Triangulating data from multiple sources made it possible to tap into some of the metacognitive strategies employed by viewers to maintain comprehension in the absence of sound. We discuss the implications within the context of the growing trend of watching subtitled videos without sound, emphasising its potential impact on cognitive processes and the viewing experience.
Assuntos
Cognição , Compreensão , Tecnologia de Rastreamento Ocular , Humanos , Feminino , Masculino , Cognição/fisiologia , Adulto , Compreensão/fisiologia , Adulto Jovem , Gravação em Vídeo , Idioma , Adolescente , Fixação Ocular/fisiologia , Som , Prazer/fisiologia , Pessoa de Meia-IdadeRESUMO
The eyes play a special role in human communications. Previous psychological studies have reported reflexive attention orienting in response to another individual's eyes during live interactions. Although robots are expected to collaborate with humans in various social situations, it remains unclear whether robot eyes have the potential to trigger attention orienting similarly to human eyes, specifically based on mental attribution. We investigated this issue in a series of experiments using a live gaze-cueing paradigm with an android. In Experiment 1, the non-predictive cue was the eyes and head of an android placed in front of human participants. Light-emitting diodes in the periphery served as target signals. The reaction times (RTs) required to localize the valid cued targets were faster than those for invalid cued targets for both types of cues. In Experiment 2, the gaze direction of the android eyes changed before the peripheral target lights appeared with or without barriers that made the targets non-visible, such that the android did not attend to them. The RTs were faster for validly cued targets only when there were no barriers. In Experiment 3, the targets were changed from lights to sounds, which the android could attend to even in the presence of barriers. The RTs to the target sounds were faster with valid cues, irrespective of the presence of barriers. These results suggest that android eyes may automatically induce attention orienting in humans based on mental state attribution.
Assuntos
Atenção , Sinais (Psicologia) , Tempo de Reação , Humanos , Atenção/fisiologia , Masculino , Feminino , Tempo de Reação/fisiologia , Adulto , Adulto Jovem , Fixação Ocular/fisiologia , Orientação/fisiologia , Robótica , OlhoRESUMO
Mental representations of the environment in infants are sparse and grow richer during their development. Anticipatory eye fixation studies show that infants aged around 7 months start to predict the goal of an observed action, e.g., an object targeted by a reaching hand. Interestingly, goal-predictive gaze shifts occur at an earlier age when the hand subsequently manipulates an object and later when an action is performed by an inanimate actor, e.g., a mechanical claw. We introduce CAPRI2 (Cognitive Action PRediction and Inference in Infants), a computational model that explains this development from a functional, algorithmic perspective. It is based on the theory that infants learn object files and events as they develop a physical reasoning system. In particular, CAPRI2 learns a generative event-predictive model, which it uses to both interpret sensory information and infer goal-directed behavior. When observing object interactions, CAPRI2 (i) interprets the unfolding interactions in terms of event-segmented dynamics, (ii) maximizes the coherence of its event interpretations, updating its internal estimates and (iii) chooses gaze behavior to minimize expected uncertainty. As a result, CAPRI2 mimics the developmental pathway of infants' goal-predictive gaze behavior. Our modeling work suggests that the involved event-predictive representations, longer-term generative model learning, and shorter-term retrospective and active inference principles constitute fundamental building blocks for the effective development of goal-predictive capacities.
Assuntos
Desenvolvimento Infantil , Cognição , Humanos , Lactente , Cognição/fisiologia , Desenvolvimento Infantil/fisiologia , Fixação Ocular/fisiologia , Feminino , Masculino , Aprendizagem/fisiologiaRESUMO
Rock climbing has propelled from niche sport to mainstream free-time activity and Olympic sport. Moreover, climbing can be studied as an example of a high-stakes perception-action task. However, understanding what constitutes an expert climber is not simple or straightforward. As a dynamic and high-risk activity, climbing requires a precise interplay between cognition, perception, and precise action execution. While prior research has predominantly focused on the movement aspect of climbing (i.e., skeletal posture and individual limb movements), recent studies have also examined the climber's visual attention and its links to their performance. To associate the climber's attention with their actions, however, has traditionally required frame-by-frame manual coding of the recorded eye-tracking videos. To overcome this challenge and automatically contextualize the analysis of eye movements in indoor climbing, we present deep learning-driven (YOLOv5) hold detection that facilitates automatic grasp recognition. To demonstrate the framework, we examined the expert climber's eye movements and egocentric perspective acquired from eye-tracking glasses (SMI and Tobii Glasses 2). Using the framework, we observed that the expert climber's grasping duration was positively correlated with total fixation duration (r = 0.807) and fixation count (r = 0.864); however, it was negatively correlated with the fixation rate (r = -0.402) and saccade rate (r = -0.344). The findings indicate the moments of cognitive processing and visual search that occurred during decision making and route prospecting. Our work contributes to research on eye-body performance and coordination in high-stakes contexts, and informs the sport science and expands the applications, e.g., in training optimization, injury prevention, and coaching.
Assuntos
Movimentos Oculares , Humanos , Movimentos Oculares/fisiologia , Tecnologia de Rastreamento Ocular , Masculino , Aprendizado Profundo , Adulto , Atenção/fisiologia , Montanhismo/fisiologia , Feminino , Fixação Ocular/fisiologiaRESUMO
As social beings, we are adept at coordinating our body movements and gaze with others. Often, when coordinating with another person, we orient ourselves to face them, as mutual gaze provides valuable cues pertaining to attention and intentions. Moreover, movement synchrony and mutual gaze are associated with prosocial outcomes, yet the perceptual consequences of these forms of coordination remain poorly understood. Across two experiments, we assessed how movement synchrony and gaze direction influence observers' perceptions of dyads. Observers' behavioural responses indicated that dyads are perceived as more socially connected and are more enjoyable to watch when moving synchronously and facing each other. Neuroimaging results showed modulation of the Action Observation and Theory of Mind networks by movement synchrony and mutual gaze, with more robust brain activity when evaluating togetherness (i.e., active and intentional collaboration) than aesthetic value (i.e., enjoyment). A fuller understanding of the consequences of movement synchrony and mutual gaze from the observer's viewpoint holds important implications for social perception, in terms of how observers intuit social relationships within dyads, and the aesthetic value derived from watching individuals moving in these ways.
Assuntos
Dança , Percepção Social , Humanos , Feminino , Masculino , Adulto , Dança/psicologia , Dança/fisiologia , Adulto Jovem , Fixação Ocular/fisiologia , Movimento/fisiologia , Atenção/fisiologia , Relações Interpessoais , Encéfalo/fisiologiaRESUMO
Humans appear to be endowed with the ability to readily share attention with interactive partners through the utilization of social direction cues, such as eye gaze and biological motion (BM). Here, we investigated the specialized brain mechanism underlying this fundamental social attention ability by incorporating different types of social (i.e., BM, gaze) and non-social (arrow) cues and combining functional magnetic resonance imaging (fMRI) with a modified central cueing paradigm. Using multi-voxel pattern analysis (MVPA), we found that although gaze- and BM-mediated attentional orienting could be decoded from neural activity in a wide range of brain areas, only the right anterior and posterior superior temporal sulcus (aSTS and pSTS) could specifically decode attentional orienting triggered by social but not non-social cues. Critically, cross-category MVPA further revealed that social attention could be decoded across BM and gaze cues in the right STS and the right superior temporal gyrus (STG). However, these regions could not decode attentional orienting across social and non-social cues. These findings together provide evidence for the existence of a specialized social attention module in the human brain, with the right STS/STG being the critical neural site dedicated to social attention.
Assuntos
Atenção , Sinais (Psicologia) , Fixação Ocular , Imageamento por Ressonância Magnética , Percepção de Movimento , Humanos , Atenção/fisiologia , Masculino , Feminino , Fixação Ocular/fisiologia , Adulto Jovem , Adulto , Percepção de Movimento/fisiologia , Mapeamento Encefálico/métodos , Percepção Social , Encéfalo/fisiologia , Encéfalo/diagnóstico por imagem , Lobo Temporal/fisiologia , Lobo Temporal/diagnóstico por imagemRESUMO
Contrary to expert readers, children learning to read have limited ability to preprocess letters in parafoveal vision. Parafoveal letters induce crowding cost: the features of neighboring letters interfere with target letter identification. We longitudinally studied the weight of parafoveal cost and benefit in two group of children (N = 42), during their first school year (Group 1) and at the end of second school year (Groupe 2). Using a novel digit-tracking method, a blurred text was presented and rendered unblurred by touching the screen, allowing the user to discover a window of visible text as the finger moved along it. We compared two conditions: (1) a large window, where crowding was enhanced by the presence of parafoveal information; (2) a small window, where crowding was suppressed by blurred parafoveal information. Finger kinematics were simultaneously recorded. We found that at the beginning of first-grade, digital fixations - brief slowing or stopping of the finger on a specific point - are significantly longer in the large compared to the small window condition, as parafoveal crowding increases text processing difficulty. This effect diminishes and disappears at the end of second-grade as reading performance improves. In the large window condition, longer digital saccades - rapid movements of the finger changing position - appear by the end of first grade suggesting that parafoveal exposure become more beneficial than harmful when children acquire basic reading skills. Our results show that in beginning readers, crowding has a cognitive cost that interfere with the speed of the learning reading process. Our findings are relevant to the field of education by showing that visual crowding in first grade should not be underestimated.
Assuntos
Dedos , Fóvea Central , Leitura , Humanos , Criança , Feminino , Masculino , Dedos/fisiologia , Fóvea Central/fisiologia , Movimento/fisiologia , Fixação Ocular/fisiologia , Aprendizagem/fisiologia , Movimentos Oculares/fisiologia , Reconhecimento Visual de Modelos/fisiologiaRESUMO
We introduce the Visual Experience Dataset (VEDB), a compilation of more than 240 hours of egocentric video combined with gaze- and head-tracking data that offer an unprecedented view of the visual world as experienced by human observers. The dataset consists of 717 sessions, recorded by 56 observers ranging from 7 to 46 years of age. This article outlines the data collection, processing, and labeling protocols undertaken to ensure a representative sample and discusses the potential sources of error or bias within the dataset. The VEDB's potential applications are vast, including improving gaze-tracking methodologies, assessing spatiotemporal image statistics, and refining deep neural networks for scene and activity recognition. The VEDB is accessible through established open science platforms and is intended to be a living dataset with plans for expansion and community contributions. It is released with an emphasis on ethical considerations, such as participant privacy and the mitigation of potential biases. By providing a dataset grounded in real-world experiences and accompanied by extensive metadata and supporting code, the authors invite the research community to use and contribute to the VEDB, facilitating a richer understanding of visual perception and behavior in naturalistic settings.
Assuntos
Movimentos Oculares , Gravação em Vídeo , Humanos , Adulto , Feminino , Adolescente , Pessoa de Meia-Idade , Masculino , Movimentos Oculares/fisiologia , Adulto Jovem , Gravação em Vídeo/métodos , Criança , Percepção Visual/fisiologia , Tecnologia de Rastreamento Ocular , Movimentos da Cabeça/fisiologia , Fixação Ocular/fisiologiaRESUMO
Real-world work environments require operators to perform multiple tasks with continual support from an automated system. Eye movement is often used as a surrogate measure of operator attention, yet conventional summary measures such as percent dwell time do not capture dynamic transitions of attention in complex visual workspace. This study analyzed eye movement data collected in a controlled a MATB-II task environment using gaze transition entropy analysis. In the study, human subjects performed a compensatory tracking task, a system monitoring task, and a communication task concurrently. The results indicate that both gaze transition entropy and stationary gaze entropy, measures of randomness in eye movements, decrease when the compensatory tracking task required more continuous monitoring. The findings imply that gaze transition entropy reflects attention allocation of operators performing dynamic operational tasks consistently.
Assuntos
Atenção , Automação , Entropia , Movimentos Oculares , Humanos , Atenção/fisiologia , Masculino , Feminino , Movimentos Oculares/fisiologia , Adulto , Fixação Ocular/fisiologia , Adulto Jovem , Tecnologia de Rastreamento Ocular , Análise e Desempenho de TarefasRESUMO
When we view the world, our eyes saccade quickly between points of interest. Even when fixating a target our eyes are not completely at rest but execute small fixational eye movements (FEMs). That vision is not blurred despite this ever-present jitter has seemingly motivated an increasingly popular theory denying the reliance of the visual system on pure spatial processing in favor of a space-to-time mechanism generated by the eye drifting across the image. Accordingly, FEMs are not detrimental but rather essential to good visibility. However, the space-to-time theory is incompatible with physiological data showing that all information is conveyed by the short neural volleys generated when the eyes land on a target, and with our faithful perception of briefly displayed objects, during which time FEMs have no effect. Another difficulty in rejecting the idea of image representation by the locations and nature of responding cells in favor of a timecode, is that somewhere, somehow, this code must be decoded into a parallel spatial one when reaching perception. Thus, in addition to the implausibility of generating meaningful responses during retinal drift, the space-to-time hypothesis calls for replacing efficient point-to-point parallel transmission with a cumbersome, delayed, space-to-time-to-space process. A novel physiological framework is presented here wherein the ability of the visual system to quickly process information is mediated by the short, powerful neural volleys generated by the landing saccades. These volleys are necessary and sufficient for normal perception without FEMs contribution. This mechanism enables our excellent perception of brief stimuli and explains that visibility is not blurred by FEMs because they do not generate useful information.
Assuntos
Fixação Ocular , Movimentos Sacádicos , Percepção Espacial , Animais , Humanos , Fixação Ocular/fisiologia , Estimulação Luminosa/métodos , Movimentos Sacádicos/fisiologia , Percepção Espacial/fisiologia , Percepção Visual/fisiologiaRESUMO
This study presents a novel multi-stage hierarchical approach to optimize key selection on virtual keyboards using eye gaze. Existing single-stage selection algorithms have difficulty with distant keys on large interfaces. The proposed technique divides the standard QWERTY keyboard into progressively smaller regions guided by eye movements, with boundary fixations first selecting halves and quarters to sequentially narrow the search area. Within each region, keys are highlighted one by one for selection. An experiment compared the multi-stage approach to single-step techniques, having participants copy text using eye gaze alone under both conditions. Metrics on selection speed, words per minute, and usability ratings were significantly improved with the hierarchical technique. Half and quarter selection times decreased over 30% on average while maintaining accuracy, with overall task completion 20% faster. Users also rated the multi-stage approach as more comfortable and easier to use. The multi-level refinement of the selection area optimized interaction efficiency for gaze-based text entry.
Assuntos
Movimentos Oculares , Tecnologia de Rastreamento Ocular , Fixação Ocular , Humanos , Fixação Ocular/fisiologia , Feminino , Masculino , Movimentos Oculares/fisiologia , Adulto , Interface Usuário-Computador , Adulto Jovem , AlgoritmosRESUMO
The foveated architecture of the human retina and the eye's mobility enables prime spatial vision, yet the interplay between photoreceptor cell topography and the constant motion of the eye during fixation remains unexplored. With in vivo foveal cone-resolved imaging and simultaneous microscopic photo stimulation, we examined visual acuity in both eyes of 16 participants while precisely recording the stimulus path on the retina. We find that resolution thresholds were correlated with the individual retina's sampling capacity, and exceeded what static sampling limits would predict by 18%, on average. The length and direction of fixational drift motion, previously thought to be primarily random, played a key role in achieving this sub-cone diameter resolution. The oculomotor system finely adjusts drift behavior towards retinal areas with higher cone densities within only a few hundred milliseconds to enhance retinal sampling.
Assuntos
Fóvea Central , Células Fotorreceptoras Retinianas Cones , Acuidade Visual , Humanos , Fóvea Central/fisiologia , Células Fotorreceptoras Retinianas Cones/fisiologia , Acuidade Visual/fisiologia , Adulto , Masculino , Feminino , Adulto Jovem , Estimulação Luminosa , Fixação Ocular/fisiologiaRESUMO
To gather visual information effectively, animals must stabilize images when they move. But closely related fruit fly species sometimes occupy strikingly diverse visual habitats. A new study shows that they have adopted gaze strategies tailored to their different visual worlds.
Assuntos
Visão Ocular , Animais , Visão Ocular/fisiologia , Percepção Visual/fisiologia , Fixação Ocular/fisiologiaRESUMO
Purpose: To propose a deep learning-based approach for predicting the most-fixated regions on optical coherence tomography (OCT) reports using eye tracking data of ophthalmologists, assisting them in finding medically salient image regions. Methods: We collected eye tracking data of ophthalmology residents, fellows, and faculty as they viewed OCT reports to detect glaucoma. We used a U-Net model as the deep learning backbone and quantized eye tracking coordinates by dividing the input report into an 11 × 11 grid. The model was trained to predict the grids on which fixations would land in unseen OCT reports. We investigated the contribution of different variables, including the viewer's level of expertise, model architecture, and number of eye gaze patterns included in training. Results: Our approach predicted most-fixated regions in OCT reports with precision of 0.723, recall of 0.562, and f1-score of 0.609. We found that using a grid-based eye tracking structure enabled efficient training and using a U-Net backbone led to the best performance. Conclusions: Our approach has the potential to assist ophthalmologists in diagnosing glaucoma by predicting the most medically salient regions on OCT reports. Our study suggests the value of eye tracking in guiding deep learning algorithms toward informative regions when experts may not be accessible. Translational Relevance: By suggesting important OCT report regions for a glaucoma diagnosis, our model could aid in medical education and serve as a precursor for self-supervised deep learning approaches to expedite early detection of irreversible vision loss owing to glaucoma.
Assuntos
Aprendizado Profundo , Tecnologia de Rastreamento Ocular , Glaucoma , Tomografia de Coerência Óptica , Humanos , Tomografia de Coerência Óptica/métodos , Glaucoma/diagnóstico , Glaucoma/diagnóstico por imagem , Oftalmologistas , Fixação Ocular/fisiologiaRESUMO
In isolated English word reading, readers have the optimal performance when their initial eye fixation is directed to the area between the beginning and word center, that is, the optimal viewing position (OVP). Thus, how well readers voluntarily direct eye gaze to this OVP during isolated word reading may be associated with reading performance. Using Eye Movement analysis with Hidden Markov Models, we discovered two representative eye movement patterns during lexical decisions through clustering, which focused at the OVP and the word center, respectively. Higher eye movement similarity to the OVP-focusing pattern predicted faster lexical decision time in addition to cognitive abilities and lexical knowledge. However, the OVP-focusing pattern was associated with longer isolated single letter naming time, suggesting conflicting visual abilities required for identifying isolated letters and multi-letter words. In contrast, in both word and pseudoword naming, although clustering did not reveal an OVP-focused pattern, higher consistency of the first fixation as measured in entropy predicted faster naming time in addition to cognitive abilities and lexical knowledge. Thus, developing a consistent eye movement pattern focusing on the OVP is essential for word orthographic processing and reading fluency. This finding has important implications for interventions for reading difficulties.
Assuntos
Movimentos Oculares , Cadeias de Markov , Leitura , Humanos , Movimentos Oculares/fisiologia , Adulto Jovem , Feminino , Masculino , Fixação Ocular/fisiologia , Adulto , Tempo de Reação/fisiologia , IdiomaRESUMO
Fixational eye movements alter the number and timing of spikes transmitted from the retina to the brain, but whether these changes enhance or degrade the retinal signal is unclear. To quantify this, we developed a Bayesian method for reconstructing natural images from the recorded spikes of hundreds of retinal ganglion cells (RGCs) in the macaque retina (male), combining a likelihood model for RGC light responses with the natural image prior implicitly embedded in an artificial neural network optimized for denoising. The method matched or surpassed the performance of previous reconstruction algorithms, and provides an interpretable framework for characterizing the retinal signal. Reconstructions were improved with artificial stimulus jitter that emulated fixational eye movements, even when the eye movement trajectory was assumed to be unknown and had to be inferred from retinal spikes. Reconstructions were degraded by small artificial perturbations of spike times, revealing more precise temporal encoding than suggested by previous studies. Finally, reconstructions were substantially degraded when derived from a model that ignored cell-to-cell interactions, indicating the importance of stimulus-evoked correlations. Thus, fixational eye movements enhance the precision of the retinal representation.
Assuntos
Movimentos Oculares , Fixação Ocular , Retina , Células Ganglionares da Retina , Animais , Células Ganglionares da Retina/fisiologia , Retina/fisiologia , Movimentos Oculares/fisiologia , Masculino , Fixação Ocular/fisiologia , Macaca mulatta , Teorema de Bayes , Algoritmos , Potenciais de Ação/fisiologia , Estimulação Luminosa , Modelos NeurológicosRESUMO
BACKGROUND: White noise stimulation has demonstrated efficacy in enhancing working memory in children with ADHD. However, its impact on other executive functions commonly affected by ADHD, such as inhibitory control, remains largely unexplored. This research aims to explore the effects of two types of white noise stimulation on oculomotor inhibitory control in children with ADHD. METHOD: Memory guided saccade (MGS) and prolonged fixation (PF) performance was compared between children with ADHD (N = 52) and typically developing controls (TDC, N = 45), during auditory and visual white noise stimulation as well as in a no noise condition. RESULTS: Neither the auditory nor the visual white noise had any beneficial effects on performance for either group. CONCLUSIONS: White noise stimulation does not appear to be beneficial for children with ADHD in tasks that target oculomotor inhibitory control. Potential explanations for this lack of noise benefit will be discussed.