Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 76
Filtrar
1.
J Sleep Res ; : e14176, 2024 Feb 25.
Artigo em Inglês | MEDLINE | ID: mdl-38404186

RESUMO

The present study aims to investigate the influence of 24-hr sleep deprivation on implicit emotion regulation using the emotional conflict task. Twenty-five healthy young adults completed a repeated-measures study protocol involving a night of at-home normal sleep control and a night of in-laboratory sleep deprivation. Prior to the experimental session, all participants wore an actigraph watch and completed the sleep diary. Following each condition, participants performed an emotional conflict task with electroencephalographic recordings. Emotional faces (fearful or happy) overlaid with words ("fear" or "happy") were used as stimuli creating congruent or incongruent trials, and participants were instructed to indicate whether the facial expression was happy or fearful. We measured the accuracy and reaction time on the emotional conflict task, as well as the mean amplitude of the P300 component of the event-related potential at CPz. At the behavioural level, sleep-deprived participants showed reduced alertness with overall longer reaction times and higher error rates. In addition, participants in the sleep deprivation condition made more errors when the current trial followed congruent trials compared with when it followed incongruent trials. At the neural level, P300 amplitude evoked under the sleep-deprived condition was significantly more positive compared with the normal sleep condition, and this effect interacted with previous-trial and current-trial congruency conditions, suggesting that participants used more attentional resources to resolve emotional conflicts when sleep deprived. Our study provided pioneering data demonstrating that sleep deprivation may impair the regulation of emotional processing in the absence of explicit instruction among emerging adults.

2.
Artigo em Inglês | MEDLINE | ID: mdl-38960910

RESUMO

Mentalizing, or theory of mind (ToM), impairments and self-referential hypermentalizing bias are well-evident in schizophrenia. However, findings compared to individuals with at-risk mental states (ARMS) are inconsistent, and investigations into the relationship between social cognitive impairments and social anxiety in the two populations are scarce. This study aimed to examine and compare these deficits in first-episode schizophrenia-spectrum disorder (FES) and ARMS, and to explore potential specific associations with neurocognition and symptomatology. Forty patients with FES, 40 individuals with ARMS, and 40 healthy controls (HC) completed clinical assessments, a battery of neurocognitive tasks, and three social cognitive tasks. The comic strip and hinting tasks were used to measure non-verbal and verbal mentalizing abilities, and the gaze perception task was employed to assess self-referential hypermentalizing bias. FES and ARMS showed comparable mentalizing impairments and self-referential hypermentalizing bias compared to HC. However, only ambiguous self-referential gaze perception (SRGP) bias remained significantly different between three groups after controlling for covariates. Findings suggested that self-referential hypermentalizing bias could be a specific deficit and may be considered a potential behavioral indicator in early-stage and prodromal psychosis. Moreover, working memory and social anxiety were related to the social cognitive impairments in ARMS, whereas higher-order executive functions and positive symptoms were associated with the impairments in FES. The current study indicates the presence of stage-specific mechanisms of mentalizing impairments and self-referential hypermentalizing bias, providing insights into the importance of personalized interventions to improve specific neurocognitive domains, social cognition, and clinical outcomes for FES and ARMS.

3.
Eur Arch Psychiatry Clin Neurosci ; 272(7): 1335-1345, 2022 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-35079856

RESUMO

Mentalizing impairment is one of the core features of schizophrenia, and bias judgement of others' gaze as self-directing is common to schizophrenia patients. In this case-control study, 30 patients with first-episode schizophrenia (FES) and 30 matched healthy controls were assigned gaze perception tasks with variable stimulus presentation times (300 ms and no time limit) to determine the presence of self-referential gaze perception (SRGP) bias. The eye movement pattern during the task were tracked and data were analysed using hidden Markov models (HMMs). The SRGP involves reporting of others' gaze intent and was used as a measurement of explicit mentalizing process. Eye movement measurement represents automated visual attention pattern and was considered as a measurement of implicit mentalizing process. The patients with FES had significantly more SRGP bias than the controls in the 300 ms condition but not in the no-time-limit condition. Social cognitive function was related to SRGP bias in the patient group. Two distinct eye movement patterns were identified: eye-focused and nose-focused. Significant group differences in eye movement patterns in the 300 ms condition were found with more controls had eye-focused pattern. Social anxiety symptoms were related to the nose-focused pattern, positive psychotic symptoms were related to the eye-focused pattern, and depressive symptoms were related to less consistent eye movement patterns. No significant relationship was found between SRGP bias and eye movement patterns. The dissociation between explicit and implicit mentalizing processes with different cognitive and symptom dimensions associated with the two processes suggests the presence of different mechanisms.


Assuntos
Mentalização , Esquizofrenia , Estudos de Casos e Controles , Movimentos Oculares , Humanos , Esquizofrenia/complicações , Percepção Social
4.
Caries Res ; 56(2): 129-137, 2022.
Artigo em Inglês | MEDLINE | ID: mdl-35398845

RESUMO

Visual attention is a significant gateway to a child's mind, and looking is one of the first behaviors young children develop. Untreated caries and the resulting poor dental aesthetics can have adverse emotional and social impacts on children's oral health-related quality of life due to its detrimental effects on self-esteem and self-concept. Therefore, we explored preschool children's eye movement patterns and visual attention to images with and without dental caries via eye movement analysis using hidden Markov models (EMHMM). We calibrated a convenience sample of 157 preschool children to the eye-tracker (Tobii Nano Pro) to ensure standardization. Consequently, each participant viewed the same standardized pictures with and without dental caries while an eye-tracking device tracked their eye movements. Subsequently, based on the sequence of viewed regions of interest (ROIs), a transition matrix was developed where the participants' previously viewed ROI informed their subsequently considered ROI. Hence, an individual's HMM was estimated from their eye movement data using a variational Bayesian approach to determine the optimal number of ROIs automatically. Consequently, this data-driven approach generated the visual task participants' most representative eye movement patterns. Preschool children exhibited two different eye movement patterns, distributed (78%) and selective (21%), which was statistically significant. Children switched between images with more similar probabilities in the distributed pattern while children remained looking at the same ROI than switching to the other ROI in the selective pattern. Nevertheless, all children exhibited an equal starting fixation on the right or left image and noticed teeth. The study findings reveal that most preschool children did not have an attentional bias to images with and without dental caries. Furthermore, only a few children selectively fixated on images with dental caries. Therefore, selective eye-movement patterns may strongly predict preschool children's sustained visual attention to dental caries. Nevertheless, future studies are essential to fully understand the developmental origins of differences in visual attention to common oral health presentations in children. Finally, EMHMM is appropriate for assessing inter-individual differences in children's visual attention.


Assuntos
Cárie Dentária , Teorema de Bayes , Pré-Escolar , Cárie Dentária/diagnóstico por imagem , Tecnologia de Rastreamento Ocular , Humanos , Saúde Bucal , Qualidade de Vida
5.
Dent Traumatol ; 38(5): 410-416, 2022 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-35460595

RESUMO

BACKGROUND/AIM: Traumatic dental injuries (TDIs) in the primary dentition may result in tooth discolouration and fractures. The aim of this child-centred study was to explore the differences between preschool children's eye movement patterns and visual attention to typical outcomes following TDIs to primary teeth. MATERIALS AND METHODS: An eye-tracker recorded 155 healthy preschool children's eye movements when they viewed clinical images of healthy teeth, tooth fractures and discolourations. The visual search pattern was analysed using the eye movement analysis with the Hidden Markov Models (EMHMM) approach and preference for the various regions of interest (ROIs). RESULTS: Two different eye movement patterns (distributed and selective) were identified (p < .05). Children with the distributed pattern shifted their fixations between the presented images, while those with the selective pattern remained focused on the same image they first saw. CONCLUSIONS: Preschool children noticed teeth. However, most of them did not have an attentional bias, implying that they did not interpret these TDI outcomes negatively. Only a few children avoided looking at images with TDIs indicating a potential negative impact. The EMHMM approach is appropriate for assessing inter-individual differences in children's visual attention to TDI outcomes.


Assuntos
Fraturas dos Dentes , Traumatismos Dentários , Pré-Escolar , Tecnologia de Rastreamento Ocular , Humanos , Dente Decíduo
6.
Sensors (Basel) ; 21(22)2021 Nov 14.
Artigo em Inglês | MEDLINE | ID: mdl-34833644

RESUMO

Mind-wandering has been shown to largely influence our learning efficiency, especially in the digital and distracting era nowadays. Detecting mind-wandering thus becomes imperative in educational scenarios. Here, we used a wearable eye-tracker to record eye movements during the sustained attention to response task. Eye movement analysis with hidden Markov models (EMHMM), which takes both spatial and temporal eye-movement information into account, was used to examine if participants' eye movement patterns can differentiate between the states of focused attention and mind-wandering. Two representative eye movement patterns were discovered through clustering using EMHMM: centralized and distributed patterns. Results showed that participants with the centralized pattern had better performance on detecting targets and rated themselves as more focused than those with the distributed pattern. This study indicates that distinct eye movement patterns are associated with different attentional states (focused attention vs. mind-wandering) and demonstrates a novel approach in using EMHMM to study attention. Moreover, this study provides a potential approach to capture the mind-wandering state in the classroom without interrupting the ongoing learning behavior.


Assuntos
Movimentos Oculares , Olho , Humanos , Aprendizagem
7.
Behav Res Methods ; 53(6): 2473-2486, 2021 12.
Artigo em Inglês | MEDLINE | ID: mdl-33929699

RESUMO

The eye movement analysis with hidden Markov models (EMHMM) method provides quantitative measures of individual differences in eye-movement pattern. However, it is limited to tasks where stimuli have the same feature layout (e.g., faces). Here we proposed to combine EMHMM with the data mining technique co-clustering to discover participant groups with consistent eye-movement patterns across stimuli for tasks involving stimuli with different feature layouts. Through applying this method to eye movements in scene perception, we discovered explorative (switching between the foreground and background information or different regions of interest) and focused (mainly looking at the foreground with less switching) eye-movement patterns among Asian participants. Higher similarity to the explorative pattern predicted better foreground object recognition performance, whereas higher similarity to the focused pattern was associated with better feature integration in the flanker task. These results have important implications for using eye tracking as a window into individual differences in cognitive abilities and styles. Thus, EMHMM with co-clustering provides quantitative assessments on eye-movement patterns across stimuli and tasks. It can be applied to many other real-life visual tasks, making a significant impact on the use of eye tracking to study cognitive behavior across disciplines.


Assuntos
Movimentos Oculares , Individualidade , Povo Asiático , Análise por Conglomerados , Humanos , Percepção Visual
8.
Cogn Emot ; 34(8): 1704-1710, 2020 12.
Artigo em Inglês | MEDLINE | ID: mdl-32552552

RESUMO

Theoretical models propose that attentional biases might account for the maintenance of social anxiety symptoms. However, previous eye-tracking studies have yielded mixed results. One explanation is that existing studies quantify eye-movements using arbitrary, experimenter-defined criteria such as time segments and regions of interests that do not capture the dynamic nature of overt visual attention. The current study adopted the Eye Movement analysis with Hidden Markov Models (EMHMM) approach for eye-movement analysis, a machine-learning, data-driven approach that can cluster people's eye-movements into different strategy groups. Sixty participants high and low in self-reported social anxiety symptoms viewed angry and neutral faces in a free-viewing task while their eye-movements were recorded. EMHMM analyses revealed novel associations between eye-movement patterns and social anxiety symptoms that were not evident with standard analytical approaches. Participants who adopted the same face-viewing strategy when viewing both angry and neutral faces showed higher social anxiety symptoms than those who transitioned between strategies when viewing angry versus neutral faces. EMHMM can offer novel insights into psychopathology-related attention processes.


Assuntos
Ansiedade/psicologia , Viés de Atenção/fisiologia , Emoções/fisiologia , Movimentos Oculares/fisiologia , Expressão Facial , Adulto , Ansiedade/fisiopatologia , Feminino , Hong Kong , Humanos , Masculino , Cadeias de Markov , Estudantes/psicologia , Estudantes/estatística & dados numéricos , Adulto Jovem
9.
Behav Res Methods ; 52(3): 1026-1043, 2020 06.
Artigo em Inglês | MEDLINE | ID: mdl-31712999

RESUMO

Here we propose the eye movement analysis with switching hidden Markov model (EMSHMM) approach to analyzing eye movement data in cognitive tasks involving cognitive state changes. We used a switching hidden Markov model (SHMM) to capture a participant's cognitive state transitions during the task, with eye movement patterns during each cognitive state being summarized using a regular HMM. We applied EMSHMM to a face preference decision-making task with two pre-assumed cognitive states-exploration and preference-biased periods-and we discovered two common eye movement patterns through clustering the cognitive state transitions. One pattern showed both a later transition from the exploration to the preference-biased cognitive state and a stronger tendency to look at the preferred stimulus at the end, and was associated with higher decision inference accuracy at the end; the other pattern entered the preference-biased cognitive state earlier, leading to earlier above-chance inference accuracy in a trial but lower inference accuracy at the end. This finding was not revealed by any other method. As compared with our previous HMM method, which assumes no cognitive state change (i.e., EMHMM), EMSHMM captured eye movement behavior in the task better, resulting in higher decision inference accuracy. Thus, EMSHMM reveals and provides quantitative measures of individual differences in cognitive behavior/style, making a significant impact on the use of eyetracking to study cognitive behavior across disciplines.


Assuntos
Movimentos Oculares , Face , Humanos , Individualidade , Cadeias de Markov , Probabilidade
10.
Cogn Affect Behav Neurosci ; 19(2): 283-295, 2019 04.
Artigo em Inglês | MEDLINE | ID: mdl-30460483

RESUMO

Sleep deprivation is suggested to impact emotion regulation, but few studies have directly examined it. This study investigated the influence of sleep deprivation on three commonly used emotion regulation strategies (distraction, reappraisal, suppression) in Gross's (1998) process model of emotion regulation. Young healthy adults were randomly assigned to a sleep deprivation group (SD; n = 26, 13 males, age = 20.0 ± 1.7) or a sleep control group (SC; n = 25, 13 males, age = 20.2 ± 1.7). Following 24-h sleep deprivation or normal nighttime sleep, participants completed an emotion regulation task, in which they naturally viewed or applied a given emotion regulation strategy towards negative pictures, with electroencephalographic (EEG) recordings. A reduction in the centroparietal late positive potential (LPP) amplitudes towards negative pictures from the naturally viewing condition to a regulated condition was calculated as an index of regulatory effects. Comparisons between the two groups indicated that sleep deprivation significantly impaired the regulatory effects of distraction and reappraisal on LPP amplitudes. Suppression did not reduce LPP amplitudes in either group. In addition, habitual sleep quality moderated the effect of sleep deprivation on subjective perception of emotional stimuli, such that sleep deprivation only made good sleepers perceive negative pictures as more unpleasant and more arousing, but it had no significant effect on poor sleepers' perception of negative pictures. Altogether, this study provides the first evidence that sleep deprivation may impair the effectiveness of applying adaptive emotion regulation strategies (distraction and reappraisal), creating potentially undesirable consequences to emotional well-being.


Assuntos
Encéfalo/fisiologia , Regulação Emocional/fisiologia , Privação do Sono/fisiopatologia , Adulto , Nível de Alerta , Eletroencefalografia , Potenciais Evocados , Feminino , Humanos , Masculino , Adulto Jovem
11.
J Sleep Res ; 28(3): e12671, 2019 06.
Artigo em Inglês | MEDLINE | ID: mdl-29493041

RESUMO

Resting-state spontaneous neural activities consume far more biological energy than stimulus-induced activities, suggesting their significance. However, existing studies of sleep loss and emotional functioning have focused on how sleep deprivation modulates stimulus-induced emotional neural activities. The current study aimed to investigate the impacts of sleep deprivation on the brain network of emotional functioning using electroencephalogram during a resting state. Two established resting-state electroencephalogram indexes (i.e. frontal alpha asymmetry and frontal theta/beta ratio) were used to reflect the functioning of the emotion regulatory neural network. Participants completed an 8-min resting-state electroencephalogram recording after a well-rested night or 24 hr sleep deprivation. The Sleep Deprivation group had a heightened ratio of the power density in theta band to beta band (theta/beta ratio) in the frontal area than the Sleep Control group, suggesting an affective approach with reduced frontal cortical regulation of subcortical drive after sleep deprivation. There was also marginally more left-lateralized frontal alpha power (left frontal alpha asymmetry) in the Sleep Deprivation group compared with the Sleep Control group. Besides, higher theta/beta ratio and more left alpha lateralization were correlated with higher sleepiness and lower vigilance. The results converged in suggesting compromised emotional regulatory processes during resting state after sleep deprivation. Our work provided the first resting-state neural evidence for compromised emotional functioning after sleep loss, highlighting the significance of examining resting-state neural activities within the affective brain network as a default functional mode in investigating the sleep-emotion relationship.


Assuntos
Eletroencefalografia/métodos , Emoções/fisiologia , Privação do Sono/fisiopatologia , Adulto , Feminino , Humanos , Masculino , Adulto Jovem
12.
J Vis ; 19(4): 10, 2019 04 01.
Artigo em Inglês | MEDLINE | ID: mdl-30952161

RESUMO

Recent research has suggested that the visual span in stimulus identification can be enlarged through perceptual learning. Since both English and music reading involve left-to-right sequential symbol processing, music-reading experience may enhance symbol identification through perceptual learning particularly in the right visual field (RVF). In contrast, as Chinese can be read in all directions, and components of Chinese characters do not consistently form a left-right structure, this hypothesized RVF enhancement effect may be limited in Chinese character identification. To test these hypotheses, here we recruited musicians and nonmusicians who read Chinese as their first language (L1) and English as their second language (L2) to identify music notes, English letters, Chinese characters, and novel symbols (Tibetan letters) presented at different eccentricities and visual field locations on the screen while maintaining central fixation. We found that in English letter identification, significantly more musicians achieved above-chance performance in the center-RVF locations than nonmusicians. This effect was not observed in Chinese character or novel symbol identification. We also found that in music note identification, musicians outperformed nonmusicians in accuracy in the center-RVF condition, consistent with the RVF enhancement effect in the visual span observed in English-letter identification. These results suggest that the modulation of music-reading experience on the visual span for stimulus identification depends on the similarities in the perceptual processes involved.


Assuntos
Povo Asiático , Compreensão/fisiologia , Idioma , Música , Reconhecimento Visual de Modelos/fisiologia , Leitura , Adulto , Feminino , Humanos , Aprendizagem , Masculino , Campos Visuais/fisiologia , Adulto Jovem
13.
Behav Res Methods ; 50(1): 362-379, 2018 02.
Artigo em Inglês | MEDLINE | ID: mdl-28409487

RESUMO

How people look at visual information reveals fundamental information about them; their interests and their states of mind. Previous studies showed that scanpath, i.e., the sequence of eye movements made by an observer exploring a visual stimulus, can be used to infer observer-related (e.g., task at hand) and stimuli-related (e.g., image semantic category) information. However, eye movements are complex signals and many of these studies rely on limited gaze descriptors and bespoke datasets. Here, we provide a turnkey method for scanpath modeling and classification. This method relies on variational hidden Markov models (HMMs) and discriminant analysis (DA). HMMs encapsulate the dynamic and individualistic dimensions of gaze behavior, allowing DA to capture systematic patterns diagnostic of a given class of observers and/or stimuli. We test our approach on two very different datasets. Firstly, we use fixations recorded while viewing 800 static natural scene images, and infer an observer-related characteristic: the task at hand. We achieve an average of 55.9% correct classification rate (chance = 33%). We show that correct classification rates positively correlate with the number of salient regions present in the stimuli. Secondly, we use eye positions recorded while viewing 15 conversational videos, and infer a stimulus-related characteristic: the presence or absence of original soundtrack. We achieve an average 81.2% correct classification rate (chance = 50%). HMMs allow to integrate bottom-up, top-down, and oculomotor influences into a single model of gaze behavior. This synergistic approach between behavior and machine learning will open new avenues for simple quantification of gazing behavior. We release SMAC with HMM, a Matlab toolbox freely available to the community under an open-source license agreement.


Assuntos
Movimentos Oculares , Aprendizado de Máquina , Cadeias de Markov , Estimulação Luminosa/métodos , Fixação Ocular , Humanos , Individualidade , Probabilidade , Análise e Desempenho de Tarefas
14.
Psychol Sci ; 25(9): 1757-67, 2014 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-25085866

RESUMO

Holistic processing and left-side bias are both behavioral markers of expert face recognition. By contrast, expert recognition of characters in Chinese orthography involves left-side bias but reduced holistic processing, although faces and Chinese characters share many visual properties. Here, we examined whether this reduction in holistic processing of Chinese characters can be better explained by writing experience than by reading experience. Compared with Chinese nonreaders, Chinese readers who had limited writing experience showed increased holistic processing, whereas Chinese readers who could write characters fluently showed reduced holistic processing. This result suggests that writing and sensorimotor experience can modulate holistic-processing effects and that the reduced holistic processing observed in expert Chinese readers may depend mostly on writing experience. However, both expert writers and writers with limited experience showed similarly stronger left-side bias than novices did in processing mirror-symmetric Chinese characters; left-side bias may therefore be a robust expertise marker for object recognition that is uninfluenced by sensorimotor experience.


Assuntos
Cognição , Escrita Manual , Reconhecimento Visual de Modelos , Leitura , Humanos , Adulto Jovem
15.
J Vis ; 14(11)2014 Sep 16.
Artigo em Inglês | MEDLINE | ID: mdl-25228627

RESUMO

We use a hidden Markov model (HMM) based approach to analyze eye movement data in face recognition. HMMs are statistical models that are specialized in handling time-series data. We conducted a face recognition task with Asian participants, and model each participant's eye movement pattern with an HMM, which summarized the participant's scan paths in face recognition with both regions of interest and the transition probabilities among them. By clustering these HMMs, we showed that participants' eye movements could be categorized into holistic or analytic patterns, demonstrating significant individual differences even within the same culture. Participants with the analytic pattern had longer response times, but did not differ significantly in recognition accuracy from those with the holistic pattern. We also found that correct and wrong recognitions were associated with distinctive eye movement patterns; the difference between the two patterns lies in the transitions rather than locations of the fixations alone.


Assuntos
Movimentos Oculares/fisiologia , Face/fisiologia , Reconhecimento Visual de Modelos/fisiologia , Reconhecimento Psicológico/fisiologia , Adolescente , Feminino , Humanos , Masculino , Cadeias de Markov , Modelos Estatísticos , Probabilidade , Adulto Jovem
16.
Top Cogn Sci ; 2024 May 23.
Artigo em Inglês | MEDLINE | ID: mdl-38781432

RESUMO

One important goal of cognitive science is to understand the mind in terms of its representational and computational capacities, where computational modeling plays an essential role in providing theoretical explanations and predictions of human behavior and mental phenomena. In my research, I have been using computational modeling, together with behavioral experiments and cognitive neuroscience methods, to investigate the information processing mechanisms underlying learning and visual cognition in terms of perceptual representation and attention strategy. In perceptual representation, I have used neural network models to understand how the split architecture in the human visual system influences visual cognition, and to examine perceptual representation development as the results of expertise. In attention strategy, I have developed the Eye Movement analysis with Hidden Markov Models method for quantifying eye movement pattern and consistency using both spatial and temporal information, which has led to novel findings across disciplines not discoverable using traditional methods. By integrating it with deep neural networks (DNN), I have developed DNN+HMM to account for eye movement strategy learning in human visual cognition. The understanding of the human mind through computational modeling also facilitates research on artificial intelligence's (AI) comparability with human cognition, which can in turn help explainable AI systems infer humans' belief on AI's operations and provide human-centered explanations to enhance human-AI interaction and mutual understanding. Together, these demonstrate the essential role of computational modeling methods in providing theoretical accounts of the human mind as well as its interaction with its environment and AI systems.

17.
Artigo em Inglês | MEDLINE | ID: mdl-38517727

RESUMO

We propose the gradient-weighted Object Detector Activation Maps (ODAM), a visual explanation technique for interpreting the predictions of object detectors. Utilizing the gradients of detector targets flowing into the intermediate feature maps, ODAM produces heat maps that show the influence of regions on the detector's decision for each predicted attribute. Compared to previous works on classification activation maps (CAM), ODAM generates instance-specific explanations rather than class-specific ones. We show that ODAM is applicable to one-stage, two-stage, and transformer-based detectors with different types of detector backbones and heads, and produces higher-quality visual explanations than the state-of-the-art in terms of both effectiveness and efficiency. We discuss two explanation tasks for object detection: 1) object specification: what is the important region for the prediction? 2) object discrimination: which object is detected? Aiming at these two aspects, we present a detailed analysis of the visual explanations of detectors and carry out extensive experiments to validate the effectiveness of the proposed ODAM. Furthermore, we investigate user trust on the explanation maps, how well the visual explanations of object detectors agrees with human explanations, as measured through human eye gaze, and whether this agreement is related with user trust. Finally, we also propose two applications, ODAM-KD and ODAM-NMS, based on these two abilities of ODAM. ODAM-KD utilizes the object specification of ODAM to generate top-down attention for key predictions and instruct the knowledge distillation of object detection. ODAM-NMS considers the location of the model's explanation for each prediction to distinguish the duplicate detected objects. A training scheme, ODAM-Train, is proposed to improve the quality on object discrimination, and help with ODAM-NMS. The code of ODAM is available: https://github.com/Cyang-Zhao/ODAM.

18.
Neural Netw ; 177: 106392, 2024 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-38788290

RESUMO

Explainable artificial intelligence (XAI) has been increasingly investigated to enhance the transparency of black-box artificial intelligence models, promoting better user understanding and trust. Developing an XAI that is faithful to models and plausible to users is both a necessity and a challenge. This work examines whether embedding human attention knowledge into saliency-based XAI methods for computer vision models could enhance their plausibility and faithfulness. Two novel XAI methods for object detection models, namely FullGrad-CAM and FullGrad-CAM++, were first developed to generate object-specific explanations by extending the current gradient-based XAI methods for image classification models. Using human attention as the objective plausibility measure, these methods achieve higher explanation plausibility. Interestingly, all current XAI methods when applied to object detection models generally produce saliency maps that are less faithful to the model than human attention maps from the same object detection task. Accordingly, human attention-guided XAI (HAG-XAI) was proposed to learn from human attention how to best combine explanatory information from the models to enhance explanation plausibility by using trainable activation functions and smoothing kernels to maximize the similarity between XAI saliency map and human attention map. The proposed XAI methods were evaluated on widely used BDD-100K, MS-COCO, and ImageNet datasets and compared with typical gradient-based and perturbation-based XAI methods. Results suggest that HAG-XAI enhanced explanation plausibility and user trust at the expense of faithfulness for image classification models, and it enhanced plausibility, faithfulness, and user trust simultaneously and outperformed existing state-of-the-art XAI methods for object detection models.


Assuntos
Inteligência Artificial , Atenção , Humanos , Atenção/fisiologia , Redes Neurais de Computação
19.
Br J Psychol ; 2024 Jun 10.
Artigo em Inglês | MEDLINE | ID: mdl-38858823

RESUMO

Explainable AI (XAI) methods provide explanations of AI models, but our understanding of how they compare with human explanations remains limited. Here, we examined human participants' attention strategies when classifying images and when explaining how they classified the images through eye-tracking and compared their attention strategies with saliency-based explanations from current XAI methods. We found that humans adopted more explorative attention strategies for the explanation task than the classification task itself. Two representative explanation strategies were identified through clustering: One involved focused visual scanning on foreground objects with more conceptual explanations, which contained more specific information for inferring class labels, whereas the other involved explorative scanning with more visual explanations, which were rated higher in effectiveness for early category learning. Interestingly, XAI saliency map explanations had the highest similarity to the explorative attention strategy in humans, and explanations highlighting discriminative features from invoking observable causality through perturbation had higher similarity to human strategies than those highlighting internal features associated with higher class score. Thus, humans use both visual and conceptual information during explanation, which serve different purposes, and XAI methods that highlight features informing observable causality match better with human explanations, potentially more accessible to users.

20.
J Cogn Neurosci ; 25(7): 998-1007, 2013 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-23448523

RESUMO

Hemispheric asymmetry in the processing of local and global features has been argued to originate from differences in frequency filtering in the two hemispheres, with little neurophysiological support. Here we test the hypothesis that this asymmetry takes place at an encoding stage beyond the sensory level, due to asymmetries in anatomical connections within each hemisphere. We use two simple encoding networks with differential connection structures as models of differential encoding in the two hemispheres based on a hypothesized generalization of neuroanatomical evidence from the auditory modality to the visual modality: The connection structure between columns is more distal in the language areas of the left hemisphere and more local in the homotopic regions in the right hemisphere. We show that both processing differences and differential frequency filtering can arise naturally in this neurocomputational model with neuroanatomically inspired differences in connection structures within the two model hemispheres, suggesting that hemispheric asymmetry in the processing of local and global features may be due to hemispheric asymmetry in connection structure rather than in frequency tuning.


Assuntos
Lateralidade Funcional/fisiologia , Modelos Neurológicos , Percepção Visual/fisiologia , Análise de Variância , Simulação por Computador , Humanos , Estimulação Luminosa
SELEÇÃO DE REFERÊNCIAS
Detalhe da pesquisa