Your browser doesn't support javascript.
loading
Emergence of integrated behaviors through direct optimization for homeostasis.
Yoshida, Naoto; Daikoku, Tatsuya; Nagai, Yukie; Kuniyoshi, Yasuo.
Affiliation
  • Yoshida N; Graduate School of Information Science and Technology, The University of Tokyo, Hongo, Bunkyo-ku, Tokyo, 113-8656, Japan; International Research Center for Neurointelligence (WPI-IRCN), Hongo, Bunkyo-ku, Tokyo, 113-0033, Japan. Electronic address: n-yoshida@isi.imi.i.u-tokyo.ac.jp.
  • Daikoku T; International Research Center for Neurointelligence (WPI-IRCN), Hongo, Bunkyo-ku, Tokyo, 113-0033, Japan.
  • Nagai Y; International Research Center for Neurointelligence (WPI-IRCN), Hongo, Bunkyo-ku, Tokyo, 113-0033, Japan; Institute for AI and Beyond, The University of Tokyo, Hongo, Bunkyo-ku, Tokyo, 113-8655, Japan.
  • Kuniyoshi Y; Graduate School of Information Science and Technology, The University of Tokyo, Hongo, Bunkyo-ku, Tokyo, 113-8656, Japan.
Neural Netw ; 177: 106379, 2024 Sep.
Article in En | MEDLINE | ID: mdl-38762941
ABSTRACT
Homeostasis is a self-regulatory process, wherein an organism maintains a specific internal physiological state. Homeostatic reinforcement learning (RL) is a framework recently proposed in computational neuroscience to explain animal behavior. Homeostatic RL organizes the behaviors of autonomous embodied agents according to the demands of the internal dynamics of their bodies, coupled with the external environment. Thus, it provides a basis for real-world autonomous agents, such as robots, to continually acquire and learn integrated behaviors for survival. However, prior studies have generally explored problems pertaining to limited size, as the agent must handle observations of such coupled dynamics. To overcome this restriction, we developed an advanced method to realize scaled-up homeostatic RL using deep RL. Furthermore, several rewards for homeostasis have been proposed in the literature. We identified that the reward definition that uses the difference in drive function yields the best results. We created two benchmark environments for homeostasis and performed a behavioral analysis. The analysis showed that the trained agents in each environment changed their behavior based on their internal physiological states. Finally, we extended our method to address vision using deep convolutional neural networks. The analysis of a trained agent revealed that it has visual saliency rooted in the survival environment and internal representations resulting from multimodal input.
Subject(s)
Key words

Full text: 1 Database: MEDLINE Main subject: Reinforcement, Psychology / Neural Networks, Computer / Homeostasis Limits: Animals / Humans Language: En Journal: Neural Netw Journal subject: NEUROLOGIA Year: 2024 Type: Article

Full text: 1 Database: MEDLINE Main subject: Reinforcement, Psychology / Neural Networks, Computer / Homeostasis Limits: Animals / Humans Language: En Journal: Neural Netw Journal subject: NEUROLOGIA Year: 2024 Type: Article