Your browser doesn't support javascript.
loading
Adversarial Feature Alignment: Avoid Catastrophic Forgetting in Incremental Task Lifelong Learning.
Yao, Xin; Huang, Tianchi; Wu, Chenglei; Zhang, Rui-Xiao; Sun, Lifeng.
Afiliação
  • Yao X; Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China yaox16@mails.tsinghua.edu.cn.
  • Huang T; Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China htc19@mails.tsinghua.edu.cn.
  • Wu C; Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China wucl18@mails.tsinghua.edu.cn.
  • Zhang RX; Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China zhangrx17@mails.tsinghua.edu.cn.
  • Sun L; Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China sunlf@tsinghua.edu.cn.
Neural Comput ; 31(11): 2266-2291, 2019 11.
Article em En | MEDLINE | ID: mdl-31525313
Humans are able to master a variety of knowledge and skills with ongoing learning. By contrast, dramatic performance degradation is observed when new tasks are added to an existing neural network model. This phenomenon, termed catastrophic forgetting, is one of the major roadblocks that prevent deep neural networks from achieving human-level artificial intelligence. Several research efforts (e.g., lifelong or continual learning algorithms) have proposed to tackle this problem. However, they either suffer from an accumulating drop in performance as the task sequence grows longer, or require storing an excessive number of model parameters for historical memory, or cannot obtain competitive performance on the new tasks. In this letter, we focus on the incremental multitask image classification scenario. Inspired by the learning process of students, who usually decompose complex tasks into easier goals, we propose an adversarial feature alignment method to avoid catastrophic forgetting. In our design, both the low-level visual features and high-level semantic features serve as soft targets and guide the training process in multiple stages, which provide sufficient supervised information of the old tasks and help to reduce forgetting. Due to the knowledge distillation and regularization phenomena, the proposed method gains even better performance than fine-tuning on the new tasks, which makes it stand out from other methods. Extensive experiments in several typical lifelong learning scenarios demonstrate that our method outperforms the state-of-the-art methods in both accuracy on new tasks and performance preservation on old tasks.
Assuntos

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Assunto principal: Encéfalo / Redes Neurais de Computação / Aprendizagem / Modelos Neurológicos Tipo de estudo: Prognostic_studies Limite: Humans Idioma: En Revista: Neural Comput Assunto da revista: INFORMATICA MEDICA Ano de publicação: 2019 Tipo de documento: Article País de afiliação: China

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Assunto principal: Encéfalo / Redes Neurais de Computação / Aprendizagem / Modelos Neurológicos Tipo de estudo: Prognostic_studies Limite: Humans Idioma: En Revista: Neural Comput Assunto da revista: INFORMATICA MEDICA Ano de publicação: 2019 Tipo de documento: Article País de afiliação: China