Spiking world model with multicompartment neurons for model-based reinforcement learning

用于基于模型的强化学习的多隔室神经元脉冲世界模型

阅读:4

Abstract

Brain-inspired spiking neural networks (SNNs) have garnered significant research attention in algorithm design and perception applications. However, their potential in the decision-making domain, particularly in model-based reinforcement learning, remains underexplored. In reinforcement learning, a world model refers to a predictive model that learns the environment's dynamics and enables agents to simulate future trajectories in a latent space, thereby improving sample efficiency and long-horizon planning. The difficulty lies in the need for spiking neurons with long-term temporal memory capabilities, as well as network optimization that can integrate and learn information for accurate predictions. The dynamic dendritic information integration mechanism of biological neurons brings us valuable insights for addressing these challenges. In this study, we propose a multicompartment neuron model capable of nonlinearly integrating information from multiple dendritic sources to dynamically process long sequential inputs. Based on this model, we construct a spiking world model (Spiking-WM), which integrates a spiking state-space model, a spiking convolutional encoder, and a fully connected spiking network for policy learning, to enable model-based deep reinforcement learning with SNNs. We evaluated our model using the DeepMind Control Suite, demonstrating that Spiking-WM outperforms existing SNN-based models and achieves performance comparable to artificial neural network-based world models employing Gated Recurrent Units. Furthermore, we assess the long-term memory capabilities of the proposed model in speech datasets, including Spiking Heidelberg Digits dataset, Texas Instruments/Massachusetts Institute of Technology Acoustic-Phonetic Continuous Speech Corpus, and LibriSpeech 100h, showing that our multicompartment neuron model surpasses other SNN-based architectures in processing long sequences.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。