- Poster presentation
- Open Access
Spike-based reinforcement learning of navigation
BMC Neuroscience volume 9, Article number: P72 (2008)
We have studied a spiking, reinforcement learning model derived from reward maximization [1, 2] where causal relations between pre-and postsynaptic activity set a synaptic eligibility trace [2, 3]. Neurons are modeled according to the "Integrate-and-Fire" model with escape noise. Synapses are binary and are modulated via the release probability. The synaptic release probability is updated when a global reward signal (such as dopamine) is received.
We have used the learning algorithm in a model of the Morris Water Maze task. The simulated rat explores the environment in random search. After only few trials the rat has learned to approach the goal from arbitrary start conditions, see Figure 1. The model features automatic generalization in state and action space due to coding by overlapping profiles of place cell and action cells .
Pfister JP, Toyoizumi T, Barber D, Gerstner W: Optimal Spike-Timing Dependent Plasticity for Precise Action Potential Firing in Supervised Learning. Neural Computation. 2006, 18 (6): 1309-1339. 10.1162/neco.2006.18.6.1318.
Florian RV: Reinforcement learning through modulation of spike-timing-dependent synaptic plasticity. Neural Computation. 2007, 19 (6): 1468-1502. 10.1162/neco.2007.19.6.1468.
Izhikevich EM: Solving the Distal Reward Problem through Linkage of STDP and Dopamine Signaling. Cerebral Cortex. 2007, 17: 2443-2452. 10.1093/cercor/bhl152.
Strösslin T, Sheynikhovich D, Chavarriaga R, Gerstner W: Robust self-localisation and navigation based on hippocampal place cells. Neural Networks. 2005, 18 (9): 1125-1140. 10.1016/j.neunet.2005.08.012.
About this article
Cite this article
Vasilaki, E., Urbanczik, R., Senn, W. et al. Spike-based reinforcement learning of navigation. BMC Neurosci 9, P72 (2008). https://doi.org/10.1186/1471-2202-9-S1-P72
- Reinforcement Learning
- Action Space
- Random Search
- Place Cell
- Maze Task