Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 30
Filtrar
Mais filtros










Base de dados
Intervalo de ano de publicação
1.
IEEE Trans Pattern Anal Mach Intell ; 45(1): 1174-1188, 2023 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-35130143

RESUMO

Appearance-based gaze estimation from RGB images provides relatively unconstrained gaze tracking from commonly available hardware. The accuracy of subject-independent models is limited partly by small intra-subject and large inter-subject variations in appearance, and partly by a latent subject-dependent bias. To improve estimation accuracy, we have previously proposed a gaze decomposition method that decomposes the gaze angle into the sum of a subject-independent gaze estimate from the image and a subject-dependent bias. Estimating the bias from images outperforms previously proposed calibration algorithms, unless the amount of calibration data is prohibitively large. This paper extends that work with a more complete characterization of the interplay between the complexity of the calibration dataset and estimation accuracy. In particular, we analyze the effect of the number of gaze targets, the number of images used per gaze target and the number of head positions in calibration data using a new NISLGaze dataset, which is well suited for analyzing these effects as it includes more diversity in head positions and orientations for each subject than other datasets. A better understanding of these factors enables low complexity high performance calibration. Our results indicate that using only a single gaze target and single head position is sufficient to achieve high quality calibration. However, it is useful to include variability in head orientation as the subject is gazing at the target. Our proposed estimator based on these studies (GEDDNet) outperforms state-of-the-art methods by more than 6.3%. One of the surprising findings of our work is that the same estimator yields the best performance both with and without calibration. This is convenient, as the estimator works well "straight out of the box," but can be improved if needed by calibration. However, this seems to violate the conventional wisdom that train and test conditions must be matched. To better understand the reasons, we provide a new theoretical analysis that specifies the conditions under which this can be expected. The dataset is available at http://nislgaze.ust.hk. Source code is available at https://github.com/HKUST-NISL/GEDDnet.

2.
Annu Int Conf IEEE Eng Med Biol Soc ; 2021: 4513-4517, 2021 11.
Artigo em Inglês | MEDLINE | ID: mdl-34892221

RESUMO

Stroke is the leading cause of adult disability. Robot-assisted rehabilitation systems show great promise for motor recovery after a stroke. In this work, we present a gazecontrolled robotic system for upper limb rehabilitation. Subjects perform a painting task in virtual reality. We designed a novel and challenging painting task to encourage motivation and engagement, as these are critical factors in treatment efficacy. Because the robotic system can be programmed to provide varying amounts of assistance or resistance to the subject, it can be applied to a wide range of patients at different phases of recovery. We describe here the system configured in two modes: resistive control and hierarchical control. The former is designed for later stages of recovery, where the patient's impaired limb has recovered some function. It can be configured to provide varying degrees of resistance by adjusting the properties of an admittance controller. The latter targets patients in more acute phases, where the impaired limb is less responsive. It provides a combination of assistive and corrective control. We pilot tested our system on 10 able-bodied subjects. Our results show that the system can provide varying degrees of resistive control, and that the integration of high level control modulated by gaze can improve engagement. These results suggest that the system may provide a more engaging environment for a wide range of rehabilitative therapies than currently available.


Assuntos
Robótica , Reabilitação do Acidente Vascular Cerebral , Acidente Vascular Cerebral , Realidade Virtual , Humanos , Extremidade Superior
3.
Annu Int Conf IEEE Eng Med Biol Soc ; 2021: 4796-4800, 2021 11.
Artigo em Inglês | MEDLINE | ID: mdl-34892283

RESUMO

Gaze-based interfaces are especially useful for people with disabilities involving the upper limbs or hands. Typically, users select from a number of options (e.g. letters or commands) displayed on a screen by gazing at the desired option. However, in some applications, e.g. gaze-based driving, it may be dangerous to direct gaze away from the environment towards a separate display. In addition, a purely gaze based interface can present a high cognitive load to users, as gaze is not normally used for selection and/or control, but rather for other purposes, such as information gathering. To address these issues, this paper presents a cost-effective multi-modal system for gaze based driving which combines appearance-based gaze estimates derived from webcam images with push button inputs that trigger command execution. This system uses an intuitive "direct interface", where users determine the direction of motion by gazing in the corresponding direction in the environment. We have implemented the system for both wheelchair control and robotic teleoperation. The use of our system should provide substantial benefits for patients with severe motor disabilities, such as ALS, by providing them with a more natural and affordable method of wheelchair control. We compare the performance of our system to the more conventional and common "indirect" system where gaze is used to select commands from a separate display, showing that our system enables faster and more efficient navigation.


Assuntos
Pessoas com Deficiência , Robótica , Cadeiras de Rodas , Mãos , Humanos , Movimento (Física)
4.
PLoS Comput Biol ; 17(5): e1008973, 2021 05.
Artigo em Inglês | MEDLINE | ID: mdl-33970912

RESUMO

Animals utilize a variety of active sensing mechanisms to perceive the world around them. Echolocating bats are an excellent model for the study of active auditory localization. The big brown bat (Eptesicus fuscus), for instance, employs active head roll movements during sonar prey tracking. The function of head rolls in sound source localization is not well understood. Here, we propose an echolocation model with multi-axis head rotation to investigate the effect of active head roll movements on sound localization performance. The model autonomously learns to align the bat's head direction towards the target. We show that a model with active head roll movements better localizes targets than a model without head rolls. Furthermore, we demonstrate that active head rolls also reduce the time required for localization in elevation. Finally, our model offers key insights to sound localization cues used by echolocating bats employing active head movements during echolocation.


Assuntos
Ecolocação/fisiologia , Movimentos da Cabeça , Localização de Som/fisiologia , Algoritmos , Animais , Quirópteros/fisiologia , Biologia Computacional/métodos
5.
IEEE Trans Neural Netw Learn Syst ; 32(5): 2066-2074, 2021 05.
Artigo em Inglês | MEDLINE | ID: mdl-32479406

RESUMO

Vision-based autonomous driving through imitation learning mimics the behavior of human drivers by mapping driver view images to driving actions. This article shows that performance can be enhanced via the use of eye gaze. Previous research has shown that observing an expert's gaze patterns can be beneficial for novice human learners. We show here that neural networks can also benefit. We trained a conditional generative adversarial network to estimate human gaze maps accurately from driver-view images. We describe two approaches to integrating gaze information into imitation networks: eye gaze as an additional input and gaze modulated dropout. Both significantly enhance generalization to unseen environments in comparison with a baseline vanilla network without gaze, but gaze-modulated dropout performs better. We evaluated performance quantitatively on both single images and in closed-loop tests, showing that gaze modulated dropout yields the lowest prediction error, the highest success rate in overtaking cars, the longest distance between infractions, lowest epistemic uncertainty, and improved data efficiency. Using Grad-CAM, we show that gaze modulated dropout enables the network to concentrate on task-relevant areas of the image.


Assuntos
Inteligência Artificial , Fixação Ocular , Redes Neurais de Computação , Atenção , Condução de Veículo , Aprendizado Profundo , Humanos , Processamento de Imagem Assistida por Computador , Aprendizado de Máquina , Reprodutibilidade dos Testes , Incerteza
6.
IEEE Trans Neural Syst Rehabil Eng ; 28(10): 2315-2324, 2020 10.
Artigo em Inglês | MEDLINE | ID: mdl-32795970

RESUMO

Eye typing is a hands-free method of human computer interaction, which is especially useful for people with upper limb disabilities. Users select a desired key by gazing at it in an image of a keyboard for a fixed dwell time. There is a tradeoff in selecting the dwell time; shorter dwell times lead to errors due to unintentional selections, while longer dwell times lead to a slow input speed. We propose to speed up eye typing while maintaining low error by dynamically adjusting the dwell time for each letter based on the past input history. More likely letters are assigned shorter dwell times. Our method is based on a probabilistic generative model of gaze, which enables us to assign dwell times using a principled model that requires only a few free parameters. We evaluate our model on both able-bodied subjects and subjects with a spinal cord injury (SCI). Compared to the standard dwell time method, we find consistent increases in typing speed in both cases. e.g., 41.8% faster typing for able-bodied subjects on a transcription task and 49.5% faster typing for SCI subjects in a chatbot task. We observed more inter-subject variability for SCI subjects.


Assuntos
Pessoas com Deficiência , Traumatismos da Medula Espinal , Teorema de Bayes , Computadores , Humanos
7.
Proc Natl Acad Sci U S A ; 117(11): 6156-6162, 2020 03 17.
Artigo em Inglês | MEDLINE | ID: mdl-32123102

RESUMO

The development of vision during the first months of life is an active process that comprises the learning of appropriate neural representations and the learning of accurate eye movements. While it has long been suspected that the two learning processes are coupled, there is still no widely accepted theoretical framework describing this joint development. Here, we propose a computational model of the development of active binocular vision to fill this gap. The model is based on a formulation of the active efficient coding theory, which proposes that eye movements as well as stimulus encoding are jointly adapted to maximize the overall coding efficiency. Under healthy conditions, the model self-calibrates to perform accurate vergence and accommodation eye movements. It exploits disparity cues to deduce the direction of defocus, which leads to coordinated vergence and accommodation responses. In a simulated anisometropic case, where the refraction power of the two eyes differs, an amblyopia-like state develops in which the foveal region of one eye is suppressed due to inputs from the other eye. After correcting for refractive errors, the model can only reach healthy performance levels if receptive fields are still plastic, in line with findings on a critical period for binocular vision development. Overall, our model offers a unifying conceptual framework for understanding the development of binocular vision.


Assuntos
Ambliopia/fisiopatologia , Olho/crescimento & desenvolvimento , Modelos Biológicos , Visão Binocular/fisiologia , Córtex Visual/crescimento & desenvolvimento , Acomodação Ocular/fisiologia , Simulação por Computador , Movimentos Oculares/fisiologia , Humanos , Aprendizagem/fisiologia , Refração Ocular/fisiologia , Disparidade Visual/fisiologia
8.
J Eye Mov Res ; 13(1)2020 May 13.
Artigo em Inglês | MEDLINE | ID: mdl-33828783

RESUMO

Previous research has shown that exposure to Japanese gardens reduces physiological measures of stress, e.g. heart rate, in both healthy subjects and dementia patients. However, the correlation between subjects' physiological responses and their visual behavior while viewing the garden has not yet been investigated. To address this, we developed a system to collect simultaneous measurements of eye gaze and three physiological indicators of autonomic nervous system activity: electrocardiogram, blood volume pulse, and galvanic skin response. We recorded healthy subjects' physiological/behavioral responses when they viewed two environments (an empty courtyard and a Japanese garden) in two ways (directly or as a projected 2D photograph). Similar to past work, we found that differences in subject's physiological responses to the two environments when viewed directly, but not as a photograph. We also found differences in their behavioral responses. We quantified subject's behavioral responses using several gaze metrics commonly considered to be measures of engagement of focus: average fixation duration, saccade amplitude, spatial entropy and gaze transition entropy. We found decrease in gaze transition entropy, the only metric that accounts for both the spatial and temporal properties of gaze, to have a weak positive correlation with decrease in heart rate. This suggests a relationship between engagement/focus and relaxation. Finally, we found gender differences: females' gaze patterns were more spatially distributed and had higher transition entropy than males.

9.
Front Neurorobot ; 13: 49, 2019.
Artigo em Inglês | MEDLINE | ID: mdl-31379548

RESUMO

We present a model for the autonomous and simultaneous learning of active binocular and motion vision. The model is based on the Active Efficient Coding (AEC) framework, a recent generalization of classic efficient coding theories to active perception. The model learns how to efficiently encode the incoming visual signals generated by an object moving in 3-D through sparse coding. Simultaneously, it learns how to produce eye movements that further improve the efficiency of the sensory coding. This learning is driven by an intrinsic motivation to maximize the system's coding efficiency. We test our approach on the humanoid robot iCub using simulations. The model demonstrates self-calibration of accurate object fixation and tracking of moving objects. Our results show that the model keeps improving until it hits physical constraints such as camera or motor resolution, or limits on its internal coding capacity. Furthermore, we show that the emerging sensory tuning properties are in line with results on disparity, motion, and motion-in-depth tuning in the visual cortex of mammals. The model suggests that vergence and tracking eye movements can be viewed as fundamentally having the same objective of maximizing the coding efficiency of the visual system and that they can be learned and calibrated jointly through AEC.

10.
Front Neurorobot ; 12: 66, 2018.
Artigo em Inglês | MEDLINE | ID: mdl-30386227

RESUMO

We propose a biologically inspired model that enables a humanoid robot to learn how to track its end effector by integrating visual and proprioceptive cues as it interacts with the environment. A key novel feature of this model is the incorporation of sensorimotor prediction, where the robot predicts the sensory consequences of its current body motion as measured by proprioceptive feedback. The robot develops the ability to perform smooth pursuit-like eye movements to track its hand, both in the presence and absence of visual input, and to track exteroceptive visual motions. Our framework makes a number of advances over past work. First, our model does not require a fiducial marker to indicate the robot hand explicitly. Second, it does not require the forward kinematics of the robot arm to be known. Third, it does not depend upon pre-defined visual feature descriptors. These are learned during interaction with the environment. We demonstrate that the use of prediction in multisensory integration enables the agent to incorporate the information from proprioceptive and visual cues better. The proposed model has properties that are qualitatively similar to the characteristics of human eye-hand coordination.

11.
Annu Int Conf IEEE Eng Med Biol Soc ; 2018: 315-318, 2018 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-30440401

RESUMO

Recent research has shown that design principles inherent in a Japanese style garden can reduce measures of stress in both healthy and dementia patients. However, it was not clear how subjects' visual interaction with the scene affected their physiological responses. To address that, we developed a novel non-invasive system to collect synchronized measurements of eye gaze and physiological indicators of sympathetic neural activity: the electrocardiogram, the blood volume pulse and the galvanic skin response, as subjects view a garden environment. We characterized the visual engagement of subjects using the average fixation duration, the saccade amplitude and the gaze transition entropy. We find a statistically significant positive correlation between gaze transition entropy and mean heart rate change. Our results suggest that the visual engagement of subjects with their environment may influence their physiological responses to it: more engagement may lead to more relaxation. Our results also highlight the importance of taking into account the detailed spatio-temporal characteristics of the gaze trajectory.


Assuntos
Demência , Jardins , Fixação Ocular , Humanos , Movimentos Sacádicos
12.
Annu Int Conf IEEE Eng Med Biol Soc ; 2018: 1550-1553, 2018 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-30440689

RESUMO

Previous studies have demonstrated that exposure to a Japanese garden is a non-pharmacological measure to improve the behavioral symptoms of elderly people with dementia, and that Japanese gardens are significantly more effective than other environments. However, it is not clear whether Japanese gardens have similar effects in the young. To address this open question, we measured the physiological responses of university students when viewing a Japanese garden, and compared them to the same students' responses when viewing a control space. We measured three physiological indicators of autonomous nervous system (ANS) activity: the electrocardiograph (ECG), the blood volume pulse (BVP) and the galvanic skin response (GSR). Our results suggest that the Japanese garden does not have as calming an effect on younger subjects as observed previously in elderly subjects. However, students did respond more positively to the Japanese garden than to an unstructured space. Ambient temperature was found to be a critical factor affecting heart rate and heart rate variability, but not other measures.


Assuntos
Resposta Galvânica da Pele , Jardins , Frequência Cardíaca , Adulto , Planejamento Ambiental , Feminino , Humanos , Japão , Masculino , Temperatura , Adulto Jovem
13.
Annu Int Conf IEEE Eng Med Biol Soc ; 2018: 2008-2011, 2018 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-30440794

RESUMO

Face recognition plays an import role in our daily lives. However, computer face recognition performance degrades dramatically with the presence of variations in illumination, head pose and occlusion. In contrast, the human brain can recognize target faces over a much wider range of conditions. In this paper, we investigate target face detection through electroencephalography (EEG). We address the problem of single-trial target-face detection in a rapid serial visual presentation (RSVP) paradigm. Whereas most previous approaches used support vector machines (SVMs), we use a convolutional neural network (CNN) to classify EEG signals when subjects view target and non-target face stimuli. The CNN outperforms the SVM algorithm, which is commonly used for event-related-potential (ERP) detection. We also compare the difference in performance when using animal stimuli. The proposed system can be potentially used in rapid face recognition system.


Assuntos
Eletroencefalografia , Reconhecimento Facial , Redes Neurais de Computação , Algoritmos , Potenciais Evocados , Humanos , Máquina de Vetores de Suporte
14.
Front Neurorobot ; 11: 58, 2017.
Artigo em Inglês | MEDLINE | ID: mdl-29163121

RESUMO

This paper investigates two types of eye movements: vergence and saccades. Vergence eye movements are responsible for bringing the images of the two eyes into correspondence, whereas saccades drive gaze to interesting regions in the scene. Control of both vergence and saccades develops during early infancy. To date, these two types of eye movements have been studied separately. Here, we propose a computational model of an active vision system that integrates these two types of eye movements. We hypothesize that incorporating a saccade strategy driven by bottom-up attention will benefit the development of vergence control. The integrated system is based on the active efficient coding framework, which describes the joint development of sensory-processing and eye movement control to jointly optimize the coding efficiency of the sensory system. In the integrated system, we propose a binocular saliency model to drive saccades based on learned binocular feature extractors, which simultaneously encode both depth and texture information. Saliency in our model also depends on the current fixation point. This extends prior work, which focused on monocular images and saliency measures that are independent of the current fixation. Our results show that the proposed saliency-driven saccades lead to better vergence performance and faster learning in the overall system than random saccades. Faster learning is significant because it indicates that the system actively selects inputs for the most effective learning. This work suggests that saliency-driven saccades provide a scaffold for the development of vergence control during infancy.

15.
Annu Int Conf IEEE Eng Med Biol Soc ; 2017: 795-798, 2017 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-29059992

RESUMO

This paper addresses the problem of estimating gaze location in the 3D environment using a remote eye tracker. Instead of relying only on data provided by the eye tracker, we investigate how to integrate gaze direction with the point-cloud-based representation of the scene provided by a Kinect sensor. The algorithm first combines the gaze vectors for the two eyes provided by the eye tracker into a single gaze vector emanating from a point in between the two eyes. The gaze target in the three dimensional environment is then identified by finding the point in the 3D point cloud that is closest to the gaze vector. Our experimental results demonstrate that the estimate of the gaze target location provided by this method is significantly better than that provided when considering gaze information alone. It is also better than two other methods for integrating point cloud information: (1) finding the 3D point closest to the gaze location as estimated by triangulating the gaze vectors from the two eyes, and (2) finding the 3D point with smallest average distance to the two gaze vectors considered individually. The proposed method has an average error of 1.7 cm in a workspace of 25 × 23 × 24 cm located at a distance of 60 cm from the user.


Assuntos
Fixação Ocular , Algoritmos , Olho
16.
Annu Int Conf IEEE Eng Med Biol Soc ; 2017: 1768-1771, 2017 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-29060230

RESUMO

Eye tracking systems are typically divided into two categories: remote and mobile. Remote systems, where the eye tracker is located near the object being viewed by the subject, have the advantage of being less intrusive, but are typically used for tracking gaze points on fixed two dimensional (2D) computer screens. Mobile systems such as eye tracking glasses, where the eye tracker are attached to the subject, are more intrusive, but are better suited for cases where subjects are viewing objects in the three dimensional (3D) environment. In this paper, we describe how remote gaze tracking systems developed for 2D computer screens can be used to track gaze points in a 3D environment. The system is non-intrusive. It compensates for small head movements by the user, so that the head need not be stabilized by a chin rest or bite bar. The system maps the 3D gaze points of the user onto 2D images from a scene camera and is also located remotely from the subject. Measurement results from this system indicate that it is able to estimate gaze points in the scene camera to within one degree over a wide range of head positions.


Assuntos
Fixação Ocular , Movimentos da Cabeça , Imageamento Tridimensional
17.
IEEE Trans Pattern Anal Mach Intell ; 39(7): 1346-1359, 2017 07 01.
Artigo em Inglês | MEDLINE | ID: mdl-27411216

RESUMO

This paper describes novel event-based spatio-temporal features called time-surfaces and how they can be used to create a hierarchical event-based pattern recognition architecture. Unlike existing hierarchical architectures for pattern recognition, the presented model relies on a time oriented approach to extract spatio-temporal features from the asynchronously acquired dynamics of a visual scene. These dynamics are acquired using biologically inspired frameless asynchronous event-driven vision sensors. Similarly to cortical structures, subsequent layers in our hierarchy extract increasingly abstract features using increasingly large spatio-temporal windows. The central concept is to use the rich temporal information provided by events to create contexts in the form of time-surfaces which represent the recent temporal activity within a local spatial neighborhood. We demonstrate that this concept can robustly be used at all stages of an event-based hierarchical model. First layer feature units operate on groups of pixels, while subsequent layer feature units operate on the output of lower level feature units. We report results on a previously published 36 class character recognition task and a four class canonical dynamic card pip task, achieving near 100 percent accuracy on each. We introduce a new seven class moving face recognition task, achieving 79 percent accuracy.This paper describes novel event-based spatio-temporal features called time-surfaces and how they can be used to create a hierarchical event-based pattern recognition architecture. Unlike existing hierarchical architectures for pattern recognition, the presented model relies on a time oriented approach to extract spatio-temporal features from the asynchronously acquired dynamics of a visual scene. These dynamics are acquired using biologically inspired frameless asynchronous event-driven vision sensors. Similarly to cortical structures, subsequent layers in our hierarchy extract increasingly abstract features using increasingly large spatio-temporal windows. The central concept is to use the rich temporal information provided by events to create contexts in the form of time-surfaces which represent the recent temporal activity within a local spatial neighborhood. We demonstrate that this concept can robustly be used at all stages of an event-based hierarchical model. First layer feature units operate on groups of pixels, while subsequent layer feature units operate on the output of lower level feature units. We report results on a previously published 36 class character recognition task and a four class canonical dynamic card pip task, achieving near 100 percent accuracy on each. We introduce a new seven class moving face recognition task, achieving 79 percent accuracy.

18.
J Vis ; 16(14): 10, 2016 11 01.
Artigo em Inglês | MEDLINE | ID: mdl-27832268

RESUMO

Optokinetic nystagmus (OKN) is an involuntary eye movement responsible for stabilizing retinal images in the presence of relative motion between an observer and the environment. Fully understanding the development of OKN requires a neurally plausible computational model that accounts for the neural development and the behavior. To date, work in this area has been limited. We propose a neurally plausible framework for the joint development of disparity and motion tuning in the visual cortex and of optokinetic and vergence eye-movement behavior. To our knowledge, this framework is the first developmental model to describe the emergence of OKN in a behaving organism. Unlike past models, which were based on scalar models of overall activity in different neural areas, our framework models the development of the detailed connectivity both from the retinal input to the visual cortex and from the visual cortex to the motor neurons. This framework accounts for the importance of the development of normal vergence control and binocular vision in achieving normal monocular OKN behaviors. Because the model includes behavior, we can simulate the same perturbations as past experiments, such as artificially induced strabismus. The proposed model agrees both qualitatively and quantitatively with a number of findings from the literature on both binocular vision and the optokinetic reflex. Finally, our model makes quantitative predictions about OKN behavior using the same methods used to characterize OKN in the experimental literature.


Assuntos
Simulação por Computador , Percepção de Movimento/fisiologia , Nistagmo Optocinético/fisiologia , Disparidade Visual/fisiologia , Visão Binocular/fisiologia , Córtex Visual/fisiologia , Humanos , Vias Visuais/fisiologia
19.
Neural Comput ; 27(7): 1496-529, 2015 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-25973550

RESUMO

Primary visual cortical complex cells are thought to serve as invariant feature detectors and to provide input to higher cortical areas. We propose a single model for learning the connectivity required by complex cells that integrates two factors that have been hypothesized to play a role in the development of invariant feature detectors: temporal slowness and sparsity. This model, the generative adaptive subspace self-organizing map (GASSOM), extends Kohonen's adaptive subspace self-organizing map (ASSOM) with a generative model of the input. Each observation is assumed to be generated by one among many nodes in the network, each being associated with a different subspace in the space of all observations. The generating nodes evolve according to a first-order Markov chain and generate inputs that lie close to the associated subspace. This model differs from prior approaches in that temporal slowness is not an externally imposed criterion to be maximized during learning but, rather, an emergent property of the model structure as it seeks a good model of the input statistics. Unlike the ASSOM, the GASSOM does not require an explicit segmentation of the input training vectors into separate episodes. This enables us to apply this model to an unlabeled naturalistic image sequence generated by a realistic eye movement model. We show that the emergence of temporal slowness within the model improves the invariance of feature detectors trained on this input.


Assuntos
Aprendizagem/fisiologia , Aprendizado de Máquina , Modelos Neurológicos , Córtex Visual/fisiologia , Percepção Visual/fisiologia , Movimentos Oculares/fisiologia , Humanos , Cadeias de Markov , Neurônios/fisiologia , Fatores de Tempo
20.
Annu Int Conf IEEE Eng Med Biol Soc ; 2015: 474-7, 2015 Aug.
Artigo em Inglês | MEDLINE | ID: mdl-26736302

RESUMO

We describe a two stage hidden Markov model based algorithm for inferring target identity in a 2D cursor control task where subjects are instructed to use a joystick to steer a cursor towards a target while avoiding obstacles. The first stage of the model converts a regularly sampled gaze trajectory into a sequence of fixations. The second stage then makes a determination of the end target in the cursor control task based on this sequence of fixations. In contrast to prior work, this two stage model allows for more accurate modelling of the natural eye gaze behavior of subjects, which in turn leads to increased accuracy and speed of target identification. This work demonstrates the importance of accurate gaze modelling, and paves the way for more natural and reliable hybrid Brain Computer Interfaces.


Assuntos
Algoritmos , Interfaces Cérebro-Computador , Desempenho Psicomotor
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...