- Poster presentation
- Open access
- Published:
More flexibility for code generation with GeNN v2.1
BMC Neuroscience volume 16, Article number: P291 (2015)
Background
GeNN (GPU enhanced Neuronal Networks) [1, 2] is a software framework that was designed to facilitate the use of GPUs (Graphics Processing Units) for the simulation of spiking neuronal networks. It is built on top of the CUDA (Common Unified Device Architecture) [3] application programming interface provided by NVIDIA Corporation and is entirely based on code generation: Users provide a compact description of a spiking neuronal network model and GeNN generates CUDA and C++ code to simulate it, also taking into account the specifics of the GPU hardware detected at compile time.
Methods
In this contribution we describe novel work on GeNN, which has transformed it to a yet more flexible tool for facilitating the use of GPUs for simulations accelerated by GPUs. The main innovations involve replacing previous fixed templates for synapse dynamics and learning models by user-definable code snippets, so allowing redefinition of virtually every dynamic element of a neural network simulation. This transition has also enabled the completion of the Brian2 to GeNN and SpineML to GeNN interfaces [4].
Results
GeNN now allows the free definition of all four, neuron dynamics, neuron threshold conditions, synapse dynamics and connection weight dynamics (learning). The desired behavior is encoded in code snippets that contain C++ compatible code that describes the operations that are necessary to complete one time step. Table 1 summarizes the available code slots and their function.
Other improvements in GeNN 2.1 include an improved CUDA block size estimation algorithm, access to pre- and post-synaptic variables in synaptic models, and a number of bug fixes.
Conclusion
GeNN has reached level of stability where it should be of increasing use to the wider computational neuroscience community, in particular with the completion of its interfaces to other simulators.
References
Nowotny T: Flexible neuronal network simulation framework using code generation for NVidia® CUDA™. BMC Neuroscience. 2011, 12 (Suppl 1): P239-
Yavuz E, Turner J, Nowotny T: Simulating spiking neural networks on massively parallel graphical processing units using a code generation approach with GeNN. BMC Neuroscience. 2014, 15 (Suppl 1): O1-
CUDA. accessed 2015-02-25, [http://www.nvidia.com/object/cuda_home_new.html]
Nowotny T, Cope AJ, Yavuz E, Stimberg M, Goodman DFM, Marshall J, Gurney K: SpineML and Brian 2.0 interfaces for using GPU enhanced Neuronal Networks (GeNN). BMC Neuroscience. 2014, 15 (Suppl 1): P148-
Acknowledgements
This work was supported by the EPSRC (Green Brain Project, grant number EP/J019690/1) and a Royal Academy of Engineering/Leverhulme Trust Fellowship.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Nowotny, T., Turner, J. & Yavuz, E. More flexibility for code generation with GeNN v2.1. BMC Neurosci 16 (Suppl 1), P291 (2015). https://doi.org/10.1186/1471-2202-16-S1-P291
Published:
DOI: https://doi.org/10.1186/1471-2202-16-S1-P291