Your browser doesn't support javascript.
loading
Reducing Catastrophic Forgetting With Associative Learning: A Lesson From Fruit Flies.
Shen, Yang; Dasgupta, Sanjoy; Navlakha, Saket.
Afiliação
  • Shen Y; Cold Spring Harbor Laboratory, Simons Center for Quantitative Biology, Cold Spring Harbor, NY 11724, U.S.A. yashen@cshl.edu.
  • Dasgupta S; Department of Computer Science and Engineering, University of California, San Diego, La Jolla, CA 92093, U.S.A. dasgupta@eng.ucsd.edu.
  • Navlakha S; Cold Spring Harbor Laboratory, Simons Center for Quantitative Biology, Cold Spring Harbor, NY 11724, U.S.A. navlakha@cshl.edu.
Neural Comput ; 35(11): 1797-1819, 2023 Oct 10.
Article em En | MEDLINE | ID: mdl-37725710
Catastrophic forgetting remains an outstanding challenge in continual learning. Recently, methods inspired by the brain, such as continual representation learning and memory replay, have been used to combat catastrophic forgetting. Associative learning (retaining associations between inputs and outputs, even after good representations are learned) plays an important function in the brain; however, its role in continual learning has not been carefully studied. Here, we identified a two-layer neural circuit in the fruit fly olfactory system that performs continual associative learning between odors and their associated valences. In the first layer, inputs (odors) are encoded using sparse, high-dimensional representations, which reduces memory interference by activating nonoverlapping populations of neurons for different odors. In the second layer, only the synapses between odor-activated neurons and the odor's associated output neuron are modified during learning; the rest of the weights are frozen to prevent unrelated memories from being overwritten. We prove theoretically that these two perceptron-like layers help reduce catastrophic forgetting compared to the original perceptron algorithm, under continual learning. We then show empirically on benchmark data sets that this simple and lightweight architecture outperforms other popular neural-inspired algorithms when also using a two-layer feedforward architecture. Overall, fruit flies evolved an efficient continual associative learning algorithm, and circuit mechanisms from neuroscience can be translated to improve machine computation.
Assuntos

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Redes Neurais de Computação / Dípteros Idioma: En Ano de publicação: 2023 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Redes Neurais de Computação / Dípteros Idioma: En Ano de publicação: 2023 Tipo de documento: Article