TY - GEN
T1 - An FPGA design framework for large-scale spiking neural networks
AU - Wang, Runchun
AU - Hamilton, Tara Julia
AU - Tapson, Jonathan
AU - Schaik, André van
PY - 2014
Y1 - 2014
N2 - We present an FPGA design framework for large-scale spiking neural networks, particularly the ones with a high-density of connections or all-to-all connections. The proposed FPGA design framework is based on a reconfigurable neural layer, which is implemented using a time-multiplexing approach to achieve up to 200,000 virtual neurons with one physical neuron using only a fraction of the hardware resources in commercial-off-the-shelf FPGAs (even entry level ones). Rather than using a mathematical computational model, the physical neuron was efficiently implemented with a conductance-based model, of which the parameters were randomised between neurons to emulate the variance in biological neurons. Besides these building blocks, the proposed time-multiplexed reconfigurable neural layer has an address buffer, which will generate a fixed random weight for each connection on the fly for incoming spikes. This structure effectively reduces the usage of memory. After presenting the architecture of the proposed neural layer, we present a network with 23 proposed neural layers, each containing 64k neurons, yielding 1.5 M neurons and 92 G synapses with a total spike throughput of 1.2T spikes/s, while running in real-time on a Virtex 6 FPGA.
AB - We present an FPGA design framework for large-scale spiking neural networks, particularly the ones with a high-density of connections or all-to-all connections. The proposed FPGA design framework is based on a reconfigurable neural layer, which is implemented using a time-multiplexing approach to achieve up to 200,000 virtual neurons with one physical neuron using only a fraction of the hardware resources in commercial-off-the-shelf FPGAs (even entry level ones). Rather than using a mathematical computational model, the physical neuron was efficiently implemented with a conductance-based model, of which the parameters were randomised between neurons to emulate the variance in biological neurons. Besides these building blocks, the proposed time-multiplexed reconfigurable neural layer has an address buffer, which will generate a fixed random weight for each connection on the fly for incoming spikes. This structure effectively reduces the usage of memory. After presenting the architecture of the proposed neural layer, we present a network with 23 proposed neural layers, each containing 64k neurons, yielding 1.5 M neurons and 92 G synapses with a total spike throughput of 1.2T spikes/s, while running in real-time on a Virtex 6 FPGA.
KW - computer simulation
KW - field programmable gate arrays
KW - multiplexing
KW - neural networks (computer science)
KW - neurons
UR - http://handle.uws.edu.au:8081/1959.7/uws:29100
UR - http://iscas2014.org/
U2 - 10.1109/ISCAS.2014.6865169
DO - 10.1109/ISCAS.2014.6865169
M3 - Conference Paper
SN - 9781479934324
SP - 457
EP - 460
BT - Proceedings 2014 IEEE International Symposium on Circuits and Systems, ISCAS 2014, Melbourne, Vic., 1-5 June 2014
PB - IEEE
T2 - IEEE International Symposium on Circuits and Systems
Y2 - 1 June 2014
ER -