Your browser doesn't support javascript.
loading
SEGAL time series classification - Stable explanations using a generative model and an adaptive weighting method for LIME.
Meng, Han; Wagner, Christian; Triguero, Isaac.
Afiliación
  • Meng H; College of Information Science and Engineering/College of Artificial Intelligence, China University of Petroleum (Beijing), Beijing, 102249, China; Computational Optimisation and Learning (COL) Lab, School of Computer Science, University of Nottingham, Nottingham, United Kingdom; The Lab for Uncertainty in Data and Decision Making (LUCID), School of Computer Science, University of Nottingham, Nottingham, United Kingdom. Electronic address: mhcupb@163.com.
  • Wagner C; The Lab for Uncertainty in Data and Decision Making (LUCID), School of Computer Science, University of Nottingham, Nottingham, United Kingdom.
  • Triguero I; Computational Optimisation and Learning (COL) Lab, School of Computer Science, University of Nottingham, Nottingham, United Kingdom; The Lab for Uncertainty in Data and Decision Making (LUCID), School of Computer Science, University of Nottingham, Nottingham, United Kingdom; DaSCI Andalusian Institute in Data Science and Computational Intelligence, Spain; Department of Computer Science and Artificial Intelligence, University of Granada, Granada, Spain.
Neural Netw ; 176: 106345, 2024 Aug.
Article en En | MEDLINE | ID: mdl-38733798
ABSTRACT
Local Interpretability Model-agnostic Explanations (LIME) is a well-known post-hoc technique for explaining black-box models. While very useful, recent research highlights challenges around the explanations generated. In particular, there is a potential lack of stability, where the explanations provided vary over repeated runs of the algorithm, casting doubt on their reliability. This paper investigates the stability of LIME when applied to multivariate time series classification. We demonstrate that the traditional methods for generating neighbours used in LIME carry a high risk of creating 'fake' neighbours, which are out-of-distribution in respect to the trained model and far away from the input to be explained. This risk is particularly pronounced for time series data because of their substantial temporal dependencies. We discuss how these out-of-distribution neighbours contribute to unstable explanations. Furthermore, LIME weights neighbours based on user-defined hyperparameters which are problem-dependent and hard to tune. We show how unsuitable hyperparameters can impact the stability of explanations. We propose a two-fold approach to address these issues. First, a generative model is employed to approximate the distribution of the training data set, from which within-distribution samples and thus meaningful neighbours can be created for LIME. Second, an adaptive weighting method is designed in which the hyperparameters are easier to tune than those of the traditional method. Experiments on real-world data sets demonstrate the effectiveness of the proposed method in providing more stable explanations using the LIME framework. In addition, in-depth discussions are provided on the reasons behind these results.
Asunto(s)
Palabras clave

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Asunto principal: Algoritmos Idioma: En Revista: Neural Netw Asunto de la revista: NEUROLOGIA Año: 2024 Tipo del documento: Article Pais de publicación: Estados Unidos

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Asunto principal: Algoritmos Idioma: En Revista: Neural Netw Asunto de la revista: NEUROLOGIA Año: 2024 Tipo del documento: Article Pais de publicación: Estados Unidos