Your browser doesn't support javascript.
loading
Human attention guided explainable artificial intelligence for computer vision models.
Liu, Guoyang; Zhang, Jindi; Chan, Antoni B; Hsiao, Janet H.
Afiliação
  • Liu G; School of Integrated Circuits, Shandong University, Jinan, China; Department of Psychology, University of Hong Kong, Pokfulam Road, Hong Kong. Electronic address: gyliu@sdu.edu.cn.
  • Zhang J; Huawei Research, Hong Kong. Electronic address: zhangjindi2@huawei.com.
  • Chan AB; Department of Computer Science, City University of Hong Kong, Kowloon Tong, Hong Kong. Electronic address: abchan@cityu.edu.hk.
  • Hsiao JH; Division of Social Science, Hong Kong University of Science and Technology, Clearwater Bay, Hong Kong; Department of Psychology, University of Hong Kong, Pokfulam Road, Hong Kong. Electronic address: jhhsiao@ust.hk.
Neural Netw ; 177: 106392, 2024 Sep.
Article em En | MEDLINE | ID: mdl-38788290
ABSTRACT
Explainable artificial intelligence (XAI) has been increasingly investigated to enhance the transparency of black-box artificial intelligence models, promoting better user understanding and trust. Developing an XAI that is faithful to models and plausible to users is both a necessity and a challenge. This work examines whether embedding human attention knowledge into saliency-based XAI methods for computer vision models could enhance their plausibility and faithfulness. Two novel XAI methods for object detection models, namely FullGrad-CAM and FullGrad-CAM++, were first developed to generate object-specific explanations by extending the current gradient-based XAI methods for image classification models. Using human attention as the objective plausibility measure, these methods achieve higher explanation plausibility. Interestingly, all current XAI methods when applied to object detection models generally produce saliency maps that are less faithful to the model than human attention maps from the same object detection task. Accordingly, human attention-guided XAI (HAG-XAI) was proposed to learn from human attention how to best combine explanatory information from the models to enhance explanation plausibility by using trainable activation functions and smoothing kernels to maximize the similarity between XAI saliency map and human attention map. The proposed XAI methods were evaluated on widely used BDD-100K, MS-COCO, and ImageNet datasets and compared with typical gradient-based and perturbation-based XAI methods. Results suggest that HAG-XAI enhanced explanation plausibility and user trust at the expense of faithfulness for image classification models, and it enhanced plausibility, faithfulness, and user trust simultaneously and outperformed existing state-of-the-art XAI methods for object detection models.
Assuntos
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Assunto principal: Atenção / Inteligência Artificial Limite: Humans Idioma: En Revista: Neural Netw / Neural netw / Neural networks Ano de publicação: 2024 Tipo de documento: Article

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Assunto principal: Atenção / Inteligência Artificial Limite: Humans Idioma: En Revista: Neural Netw / Neural netw / Neural networks Ano de publicação: 2024 Tipo de documento: Article