TY - JOUR
T1 - A scalable implementation of the recursive least-squares algorithm for training spiking neural networks
AU - Arthur, Benjamin J.
AU - Kim, Christopher M.
AU - Chen, Susu
AU - Preibisch, Stephan
AU - Darshan, Ran
N1 - Publisher Copyright:
Copyright © 2023 Arthur, Kim, Chen, Preibisch and Darshan.
PY - 2023
Y1 - 2023
N2 - Training spiking recurrent neural networks on neuronal recordings or behavioral tasks has become a popular way to study computations performed by the nervous system. As the size and complexity of neural recordings increase, there is a need for efficient algorithms that can train models in a short period of time using minimal resources. We present optimized CPU and GPU implementations of the recursive least-squares algorithm in spiking neural networks. The GPU implementation can train networks of one million neurons, with 100 million plastic synapses and a billion static synapses, about 1,000 times faster than an unoptimized reference CPU implementation. We demonstrate the code's utility by training a network, in less than an hour, to reproduce the activity of > 66, 000 recorded neurons of a mouse performing a decision-making task. The fast implementation enables a more interactive in-silico study of the dynamics and connectivity underlying multi-area computations. It also admits the possibility to train models as in-vivo experiments are being conducted, thus closing the loop between modeling and experiments.
AB - Training spiking recurrent neural networks on neuronal recordings or behavioral tasks has become a popular way to study computations performed by the nervous system. As the size and complexity of neural recordings increase, there is a need for efficient algorithms that can train models in a short period of time using minimal resources. We present optimized CPU and GPU implementations of the recursive least-squares algorithm in spiking neural networks. The GPU implementation can train networks of one million neurons, with 100 million plastic synapses and a billion static synapses, about 1,000 times faster than an unoptimized reference CPU implementation. We demonstrate the code's utility by training a network, in less than an hour, to reproduce the activity of > 66, 000 recorded neurons of a mouse performing a decision-making task. The fast implementation enables a more interactive in-silico study of the dynamics and connectivity underlying multi-area computations. It also admits the possibility to train models as in-vivo experiments are being conducted, thus closing the loop between modeling and experiments.
KW - Neuropixels dense silicon probe
KW - balanced networks
KW - dynamical system
KW - excitation-inhibition
KW - integrate and fire neuron
UR - http://www.scopus.com/inward/record.url?scp=85164794498&partnerID=8YFLogxK
U2 - 10.3389/fninf.2023.1099510
DO - 10.3389/fninf.2023.1099510
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
C2 - 37441157
AN - SCOPUS:85164794498
SN - 1662-5196
VL - 17
JO - Frontiers in Neuroinformatics
JF - Frontiers in Neuroinformatics
M1 - 1099510
ER -