Your browser doesn't support javascript.
loading
Tf-GCZSL: Task-free generalized continual zero-shot learning.
Gautam, Chandan; Parameswaran, Sethupathy; Mishra, Ashish; Sundaram, Suresh.
Afiliación
  • Gautam C; Institute for Infocomm Research (I2R), A*STAR, Singapore. Electronic address: gautamc@i2r.a-star.edu.sg.
  • Parameswaran S; Indian Institute of Science, Bangalore, India. Electronic address: sethupathyp@iisc.ac.in.
  • Mishra A; Indian Institute of Technology Madras, India. Electronic address: mishra@cse.iitm.ac.in.
  • Sundaram S; Indian Institute of Science, Bangalore, India. Electronic address: vssuresh@iisc.ac.in.
Neural Netw ; 155: 487-497, 2022 Nov.
Article en En | MEDLINE | ID: mdl-36162233
ABSTRACT
Learning continually from a stream of training data or tasks with an ability to learn the unseen classes using a zero-shot learning framework is gaining attention in the literature. It is referred to as continual zero-shot learning (CZSL). Existing CZSL requires clear task-boundary information during training which is not practically feasible. This paper proposes a task-free generalized CZSL (Tf-GCZSL) method with short-term/long-term memory to overcome the requirement of task-boundary in training. A variational autoencoder (VAE) handles the fundamental ZSL tasks. The short-term and long-term memory help to overcome the condition of the task boundary in the CZSL framework. Further, the proposed Tf-GCZSL method combines the concept of experience replay with dark knowledge distillation and regularization to overcome the catastrophic forgetting issues in a continual learning framework. Finally, the Tf-GCZSL uses a fully connected classifier developed using the synthetic features generated at the latent space of the VAE. The performance of the proposed Tf-GCZSL is evaluated in the existing task-agnostic prediction setting and the proposed task-free setting for the generalized CZSL over the five ZSL benchmark datasets. The results clearly indicate that the proposed Tf-GCZSL improves the prediction at least by 12%, 1%, 3%, 4%, and 3% over existing state-of-the-art and baseline methods for CUB, aPY, AWA1, AWA2, and SUN datasets, respectively in both settings (task-agnostic prediction and task-free learning). The source code is available at https//github.com/Chandan-IITI/Tf-GCZSL.
Asunto(s)
Palabras clave

Texto completo: 1 Colección: 01-internacional Banco de datos: MEDLINE Asunto principal: Aprendizaje Automático / Aprendizaje Idioma: En Revista: Neural Netw Asunto de la revista: NEUROLOGIA Año: 2022 Tipo del documento: Article

Texto completo: 1 Colección: 01-internacional Banco de datos: MEDLINE Asunto principal: Aprendizaje Automático / Aprendizaje Idioma: En Revista: Neural Netw Asunto de la revista: NEUROLOGIA Año: 2022 Tipo del documento: Article