Six Networks On A Universal Neuromorphic Computing Substrate
Six Networks On A Universal Neuromorphic Computing Substrate
* Correspondence:
Thomas Pfeil
Universitt Heidelberg
Kirchhoff-Institute for Physics
Im Neuenheimer Feld 227
69120 Heidelberg, Germany
tel: +49-6221-549813
[email protected]
These authors contributed equally to this work. See acknowledgements for details.
Abstract
In this study, we present a highly configurable neuromorphic computing substrate and use it for emulating
several types of neural networks. At the heart of this
system lies a mixed-signal chip, with analog implementations of neurons and synapses and digital transmission
of action potentials. Major advantages of this emulation
device, which has been explicitly designed as a universal
neural network emulator, are its inherent parallelism, a
high speed-up factor and low power consumption compared to conventional computers. Its configurability allows the realization of almost arbitrary network topologies and the use of widely varied neuronal and synaptic
parameters. Fixed-pattern noise inherent to analog circuitry is reduced by calibration routines. An integrated
development environment allows neuroscientists to operate the device without any prior knowledge of neuromorphic circuit design. As a showcase for the capabilities of
the system, we describe the successful emulation of six
different neural networks which cover a broad spectrum
of both structure and functionality.
Keywords: accelerated neuromorphic hardware
system, universal computing substrate, highly
configurable, mixed-signal VLSI, spiking neural networks, winner-take-all, classifier, cortical
model.
Pfeil et al.
1.
INTRODUCTION
By nature, computational neuroscience has a high demand for powerful and efficient devices for simulating
neural network models. In contrast to conventional
general-purpose machines based on a von-Neumann architecture, neuromorphic systems are, in a rather broad
definition, a class of devices which implement particular
features of biological neural networks in their physical
circuit layout (Mead, 1989; Indiveri et al., 2009; Renaud
et al., 2010). In order to discern more easily between
computational substrates, the term emulation is generally used when referring to neural networks running on a
neuromorphic back-end.
Several aspects motivate the neuromorphic approach.
Particularly notable, in this context, are emulation speed
and power efficiency. The former is enabled by the arguably most characteristic feature of neuromorphic devices: due to the fact that individual neural network components (essentially neurons and synapses) are physically
implemented in silico, parallelism is inherent. Due to this
parallelism, scaling of emulated network models does not
imply slowdown, as is usually the case for conventional
machines. The hard upper bound in network size (given
by the number of available components on the neuromorphic device) can be broken by scaling of the devices themselves, e.g., by wafer-scale integration (Schemmel et al.,
2010) or massively interconnected chips (Merolla et al.,
2011). Furthermore, dedicated circuits mimicking neural
properties, either digital (Merolla et al., 2011) or mixedsignal (Indiveri et al., 2006; Brderle et al., 2011), reduce
power consumption compared to conventional computing
architectures. Emulations can be further accelerated by
scaling down time constants compared to biology, which is
enabled by deep submicron technology (Schemmel et al.,
2006, 2010; Brderle et al., 2011).
However, in contrast to the unlimited model flexibility
offered by conventional simulation, the network topology
and parameter space of neuromorphic systems are often
dedicated for predefined applications and therefore rather
restricted (e.g., Serrano-Gotarredona et al., 2006; Merolla
& Boahen, 2006; Akay, 2007; Chicca et al., 2007). Enlarging the configuration space always comes at the cost of
hardware resources by occupying additional chip area and
increasing power consumption. Consequently, the maximum network size is reduced, or the configurability of
one aspect is decreased by increasing the configurability
of another. Still, configurability costs can be counterbalanced by decreasing precision. This could concern the
size of integration time steps (Merolla et al., 2011), the
granularity of particular parameters (Pfeil et al., 2012) or
fixed-pattern noise affecting various network components
(Neftci & Indiveri, 2010; Brderle et al., 2011). At least
the latter can be, to some extent, moderated through
On Spikey (Figure 1), a VLSI version of the standard leaky integrate-and-fire (LIF) neuron model with
conductance-based synapses is implemented (Dayan &
Abbott, 2001):
Cm
X
dVm
= gl (Vm El ) +
gi (Vm Ei )
dt
i
(1)
For its hardware implementation see Figure 1 and Schemmel et al. (2006).
Pfeil et al.
mux
Line
Driver
(A)
ext
int
exc/inh
trise tfall
Synapse
(B)
weight
RAM
Eexc
Neuron
(C)
Einh
gl
El
Vm
Vth
Cm
Vreset
FIGURE 1: Microphotograph of the Spikey chip (fabricated in a 180 nm CMOS process with die size 5 5 mm2 ). Each of its 384
neurons can be arbitrarily connected to any other neuron. In the following, we give a short overview of the technical implementation
of neural networks on the Spikey chip. (A) Within the synapse array 256 synapse line drivers convert incoming digital spikes (blue)
into a linear voltage ramp (red). They are individually driven by either another on-chip neuron (int), e.g., the one depicted in (C),
or an external spike source (ext). (B) Within the synapse, depending on its individually configurable weight wi , the linear voltage
ramp (red) is then translated into a current pulse (green) with exponential onset and decay. These postsynaptic pulses are sent to
the neuron via the excitatory (exc) and inhibitory (inh) input line, shared by all synapses in that array column. (C) Upon reaching
the neuron circuit, the total current on both input lines is converted into conductances, respectively. If the membrane potential Vm
crosses the firing threshold Vth , a digital pulse (blue) is generated, which can be recorded and fed back into the synapse array. After
any spike, Vm is set to Vreset for a refractory time period of refrac . Neuron and synapse line driver parameters can be configured as
summarized in Table 1.
(2)
SYSTEM ENVIRONMENT
Pfeil et al.
Network Module
Spikey Chip
Neuromorphic Network
Sequencer
DAC/ADC
(FPGA)
RAM
Host Computer
PyNN
(neuronal network
modeling language)
Control
Software
CONFIGURABILITY
In order to facilitate the emulation of network models inspired by biological neural structures, it is essential
to support the implementation of different (cortical) neuron types. From a mathematical perspective, this can be
achieved by varying the appropriate parameters of the
implemented neuron model (Equation 1).
To this end, the Spikey chip provides 2969 different
analog parameters (Table 1) stored on current memory
cells that are continuously refreshed from a digital onchip memory. Most of these cells deliver individual parameters for each neuron (or synapse line driver), e.g.,
leakage conductances gl . Due to the size of the currentvoltage conversion circuitry it was not possible to provide
individual voltage parameters, such as, e.g., El , Eexc and
Einh , for each neuron. As a consequence, groups of 96
neurons share most of these voltage parameters. Parameters that can not be controlled individually are delivered
by global current memory cells.
In addition to the possibility of controlling analog parameters, the Spikey chip also offers an almost arbitrary
configurability of the network topology. As illustrated in
Figure 1, the fully configurable synapse array allows connections from synapse line drivers (located alongside the
array) to arbitrary neurons (located below the array) via
synapses whose weights can be set individually with a 4bit resolution. This limits the maximum input count to
256 synapses per neuron, up to 192 synapses from on-chip
neurons and up to 256 from external spike sources.
2.4.
CALIBRATION
Pfeil et al.
Scope
Name
Type Description
refrac
El
Einh
Eexc
Vth
Vreset
in
in
in
sn
sn
sn
sn
sn
Two digital configuration bits activating the neuron and readout of its membrane voltage
Bias current for neuron leakage circuit
Bias current controlling neuron refractory time
Leakage reversal potential
Inhibitory reversal potential
Excitatory reversal potential
Firing threshold voltage
Reset potential
Synapse line
drivers (B)
n/a
n/a
trise , tfall
gimax
il
il
il
il
Synapses (B)
is
STP
related (C)
n/a
USE
n/a
rec , facil
I
R
il
il
sl
sl
sl
sl
STDP
related (D)
n/a
A+/n/a
STDP
il
sl
sl
g
Bias current controlling delay for presynaptic correlation pulse (for calibration purposes)
Two voltages dimensioning charge accumulation per (anti-)causal correlation measurement
Two threshold voltages for detection of relevant (anti-)causal correlation
Voltage controlling STDP time constants
n/a
gl
Neuron
circuits (A)
TABLE 1: List of analog current and voltage parameters as well as digital configuration bits. Each with corresponding model
parameter names, excluding technical parameters that are only relevant for correctly biasing analog support circuitry or controlling
digital chip functionality. Electronic parameters that have no direct translation to model parameters are denoted n/a. The membrane
capacitance is fixed and identical for all neuron circuits (Cm = 0.2 nF in biological value domain). Parameter types: (i) controllable
for each corresponding circuit: 192 for neuron circuits (denoted with subscript n), 256 for synapse line drivers (denoted with subscript
l), 49152 for synapses (denoted with subscript s), (s) two values, shared for all even/odd neuron circuits or synapse line drivers,
respectively, (g) global, one value for all corresponding circuits on the chip. All numbers refer to circuits associated to one synapse
array and are doubled for the whole chip. For technical reasons, the current revision of the chip only allows usage of one synapse
array of the chip. Therefore, all experiments presented in this paper are limited to a maximum of 192 neurons. For parameters
denoted by (A) see Equation 1 and Schemmel et al. (2006), for (B) see Figure 1, Equation 2 and Dayan & Abbott (2001), for (C)
see Schemmel et al. (2007) and for (D) see Schemmel et al. (2006) and Pfeil et al. (2012).
as to determine the dynamic range of all model parameters (e.g., Brderle & Mller et al., 2009).
To facilitate modeling and provide repeatability of experiments on arbitrary Spikey chips, it is essential to minimize these effects by calibration routines. Many calibrations have directly corresponding biological model parameters, e.g., membrane time constants (described in the
following), firing thresholds, synaptic efficacies or PSP
shapes. Others have no equivalents, like compensations
for shared parameters or workarounds of defects (e.g., Kaplan et al., 2009; Bill et al., 2010; Pfeil et al., 2012). In
general, calibration results are used to improve the mapping between biological input parameters and the corresponding target hardware voltages and currents, as well
While the calibration of most parameters is rather technical, but straightforward (e.g., all neuron voltage parameters), some require more elaborate techniques. These include the calibration of m , STP as well as synapse line
drivers, as we describe later for individual network models. The membrane time constant m = Cm /gl differs
from neuron to neuron mostly due to variations in the
leakage conductance gl . However, gl is independently adjustable for every neuron. Because this conductance is
not directly measurable, an indirect calibration method
is employed. To this end, the threshold potential is set
below the resting potential. Following each spike, the
membrane potential is clamped to Vreset for an absolute
Pfeil et al.
Network Topology
Pfeil et al.
(Background)
150
RS
Stim.
neuron index
FS
alast
C
50
(mV)
2
1
2 4 6 8
(ms)
5 10 15 20 25 30 35
(ms)
Vm
100
60
70
80
1000
1025
1050
time (ms)
1075
FIGURE 3: (A) Synfire chain with feedforward inhibition. The background is only utilized in the original model, where it is
implemented as random Gaussian current. For the presented hardware implementation it has been omitted due to network size
constraints. As compensation for missing background stimuli, the resting potential was increased to ensure a comparable excitability
of the neurons. (B) Hardware emulation. Top: Raster plot of pulse packet propagation 1000 ms after initial stimulus. Spikes from
RS groups are shown in red and spikes from FS groups are denoted by blue color and background. Bottom: Membrane potential of
the first neuron in the fourth RS group, which is denoted by a dashed horizontal line. The cycle duration is approximately 20 ms. (C)
State space generated with software simulations of the original model. The position of each marker indicates the (, a) parameters
of the stimulus while the color encodes the activity in the RS population of the third synfire group. Lack of activity is indicated with
a cross. The evolution of the pulse packet parameters is shown for three selected cases by a series of arrows. Activity either stably
propagates with fixed point (, a) = (0.1 ms, 1) or extinguishes with fixed point (, a) = (0 ms, 0). (D) Same as (C), but emulated
on the FACETS chip-based system. The fixed point of stable propagation is located at (, a) = (0.3 ms, 1).
3.1.2.
Hardware Emulation
Pfeil et al.
Network Topology
BRNs consist of an inhibitory and excitatory population of neurons, both stimulated by two populations of
Poisson processes mimicking background activity. Additionally, both neuron populations project onto themselves
as well as the other. All projections are realized with random and sparse connections of probability p. In order to
obtain an asynchronous irregular network state, synaptic
weights for inhibitory connections are chosen four times
larger than those for excitatory ones. In contrast to the
original implementation using 12500 neurons, we scaled
this network by a factor of 100 while preserving its firing
behavior.
If single cells fire irregularly, the coefficient of variation
CV =
T
T
(3)
Hardware Emulation
In addition to standard calibration routines (Section 2.4), we have calibrated the chip explicitly for the
BRN shown in Figure 4A. In the first of two steps, excitatory and inhibitory synapse line drivers were calibrated sequentially towards equal strength, respectively,
but with inhibition four times stronger than excitation.
To this end, all available neurons received spiking activity from a single synapse line driver, thereby averaging
out neuron-to-neuron variations. The shape of synaptic
conductances (specifically tfall and gimax ) were adjusted to
obtain a target mean firing rate of 10 Hz over all neurons.
Similarly, each driver was calibrated for its inhibitory operation mode. All neurons were strongly stimulated by
an additional driver with its excitatory mode already calibrated, and again the shape of conductances, this time
for inhibition, was adjusted to obtain the target rate.
Untouched by this prior calibration towards a target
mean rate, neuron excitability still varied between neurons and was calibrated consecutively for each neuron in a
second calibration step. For this, all neurons of the BRN
were used to stimulate a single neuron with a total firing
rate that was uniformly distributed among all inputs and
equal to the estimated firing rate of the final network implementation. Subsequently, all afferent synaptic weights
to this neuron were scaled in order to adapt its firing rate
to the target rate. To further stabilize the network, efferent connections of the excitatory neuron population were
modeled as short-term depressing.
Figure 4B and C show recordings of a BRN emulation
on a calibrated chip with neurons firing irregularly and
asynchronously. Note that CV 1 does not necessarily
guarantee an exponential interspike interval distribution
and even less Poisson firing. However, neurons within the
BRN clearly exhibit irregular firing (Figure 4B and D).
A simulation of the same network topology and stimulus using software tools produced similar results. Synaptic weights were not known for the hardware emulation,
but defined by the target firing rates using the above calibration. A translation to biological parameters is possible, but would have required further measurements and
was not of further interest in this context. Instead, for
software simulations, the synaptic weight for excitatory
connections were chosen to fit the mean firing rate of the
hardware emulation (approx. 9 Hz). Then, the weight of
inhibitory connections were chosen to preserve the ratio
between inhibitory and excitatory weights.
Pfeil et al.
Nq
100
neuron index
100
75
50
25
1
E
50
55
60
65
70
75
13.0 13.2 13.4 13.6 13.8 14.0
t [s]
Vm,SW [mV]
Np
Ne
D 125
neuron index
Ni
B 125
Vm,HW [mV]
75
50
25
1
50
55
60
65
13.0 13.2 13.4 13.6 13.8 14.0
t [s]
FIGURE 4: Network topology of a balanced random network. (A) Populations consisting of Ne = 100 excitatory and Ni = 25
inhibitory neurons (gray circles), respectively, are stimulated by populations of Poisson sources (black circles). We use Np = 100
independent sources for excitation and Nq = 25 for inhibition. Arrows denote projections between these populations with connection
probabilities p = 0.1, with solid lines for excitatory and dotted lines for inhibitory connections. Dot and dash lines are indicating
excitatory projections with short-term depression. (B) Raster plot for a hardware emulation (HW) on the Spikey chip. Populations
of excitatory and inhibitory neurons are depicted with white and gray background, respectively. Note that for clarity only the time
interval [13 s..14 s] of a 20 s emulation is shown. For the full 20 s emulation, we have measured CV = 1.02 0.16 (mean over all
neurons) and CC = 0.0140.019 (mean over 1000 random chosen pairs of neurons), respectively. (C) Recorded membrane potential
of an arbitrary excitatory neuron (neuron index 3, highlighted in red in the above raster plot). (D), (E) Same network topology and
stimulus as in (B) and (C), but simulated with software (SW), resulting in CV = 0.96 0.09 and CC = 0.010 0.017.
Membrane dynamics of single neurons within the network are comparable between hardware emulations and
software simulations (Figure 4C and E). Evidently, spike
times differ between the two approaches due to various
hardware noise sources (Section 2.4). However, we observe that these phenomena have no effect in large populations of neurons and result in firing statistics comparable to software simulations (compare Figure 4B and D).
The ability to reproduce these statistics is highly relevant in the context of cortical models which rely on asynchronous irregular firing activity for information processing (e.g., van Vreeswijk & Sompolinsky, 1996).
3.3.
WINNER-TAKE-ALL NETWORK
3.3.1.
Network Topology
Hardware Emulation
Hardware emulations of the WTA network show comparable discrimination performance to software simulations (Figure 5B and C). Figure 5D and E depict activity
profiles of the neuron layers illustrating recurrent excitation. The asymmetry in case of hardware neurons expresses the inhomogeneity of their excitability (Section 2).
In case of hardware emulations, each stimulus was distributed and hence averaged over 5 synapse line drivers
in order to equalize stimulation strength among neurons.
For both back-ends, inhibitory weights were chosen four
times stronger than excitatory ones (using the synapse
line driver calibration of Section 3.2), but their absolute values differed. However, the hardware system shows
Pfeil et al.
C 700
B 200
100
Inhibitory Pool
ext
Stim. Stim. Stim.
rSW [Hz]
50
00
60
50
40
30
20
10
026
600
500
400
300
200
100
00
100
E 140
120
100
80
60
40
20
026
50
rtot,HW [Hz]
150
20
40
r2
60
[Hz]
80
20
40
r2
60
[Hz]
80
100
rHW [Hz]
rec
rtot,SW [Hz]
44
32
38
neuron index
44
32
38
neuron index
50
FIGURE 5: (A) Topology of a soft winner-take-all network. Recurrently connected excitatory neurons are located on a ring-shaped
layer (50 neurons; gray circles), and project bidirectionally to a common inhibitory pool (16 neurons; upper black ellipse). Each of
the excitatory neurons projects to its neighboring neurons with synaptic weights decreasing with distance on the ring. More precisely,
synaptic weights are a function of distance following a Gaussian profile with its mean at the presynaptic neuron, whereas the
standard deviation is expressed by the distance between neurons (for recurrent connections between excitatory neurons: rec = 5
neurons; blue). Similarly, weights of external stimulation also follow a Gaussian profile with mean ext and standard deviation
ext = 3 neurons (red). Each neuron located within a distance of 3ext is stimulated by an external Poisson spike sources firing
at 5 Hz (lower black ellipse). Solid arrows denote excitatory and dotted arrows inhibitory connections. (B) Software simulation
(SW). One side of the ring (ext = neuron index 13; black) is stimulated with r1 = 50 Hz while the firing rate r2 of the stimulus
on the other side (ext = neuron index 38; gray) is varied. The total output firing rate rtot (mean over 5 runs at 2 s duration with
different random numbers each) for each half of the ring is plotted against r2 . (C) Same as (B), but emulated on Spikey (HW).
(D) Firing rate distribution over neuron indices for r2 = 25 Hz (black), 50 Hz (dark gray) and 75 Hz (light gray). (E) Same as (D),
but emulated on Spikey .
Network Topology
From a structural perspective, the most prominent feature of the Layer 2/3 Attractor Memory Network is its
modularity. Faithful to its biological archetype, it implements a set of cortical hypercolumns, which are in
turn subdivided into multiple minicolumns (Figure 6A).
Each minicolumn consists of three cell populations: excitatory pyramidal cells, inhibitory basket cells and inhibitory RSNP (regular spiking non-pyramidal) cells.
Attractor dynamics arise from the synaptic connectivity on two levels. Within a hypercolumn, the basket cell
160
140
120
100
80
60
40
20
0
1000 2000 3000 4000 5000
time [ms]
RSNP
MC1
pyramidal
MC2
0
E
35
30
25
20
15
10
5 0.0 0.2 0.4 0.6 0.8 1.0
time / dwell time
2
1
basket
neuron #
pattern number
MC1 MC2
pattern number
2
1
0
68
70
72
74
76
78 1.0 0.5 0.0 0.5 1.0 1.5 2.0 2.5
time / dwell time
160
140
120
100
80
60
40
20
0
1000 2000 3000 4000 5000
time [ms]
neuron #
Pfeil et al.
1.0
0.9
0.8
0.7
0.6
0.5
0.4
0.30.0
2.5
FIGURE 6: (A) Schematic of the cortical layer 2/3 attractor memory network. Two hypercolumns, each containing two minicolumns,
are shown. For better readability, only connections that are active within an active pattern are depicted. See text for details. (B)
Software simulation of spiking activity in the cortical attractor network model scaled down to 192 neurons (only pyramidal and RSNP
cells shown, basket cells spike almost continously). Minicolumns belonging to the same pattern are grouped together. The broad
stripes of activity are generated by pyramidal cells in active attractors. The interlaced narrow stripes of activity represent pairs of
RSNP cells, which spike when their home minicolumn is inhibited by other active patterns. (C) Same as B, but on hardware. The
raster plot is noisier and the duration of attractors (dwell time) are less stable than in software due to fixed-pattern noise on neuron
and synapse circuits. For better readability, active states are underlied in grey in B and C. (D) Average firing rate of pyramidal cells
on the Spikey chip inside active patterns. (E) Average membrane potential of pyramidal cells on the Spikey chip inside and outside
active patterns. To allow averaging over multiple active periods of varying lengths, all attractor dwell times have been normalized to
1 in D and E. (F) Pattern completion on the Spikey chip. Average values (from multiple runs) depicted in blue, with the standard
deviation shown in red. From a relatively equilibrated state where all patterns take turns in being active, additional stimulation
(see text) of only a subset of neurons from a given attractor activates the full pattern and enables it to dominate over the other
two. The pattern does not remain active indefinitely due to short-term depression in excitatory synapses, thereby still allowing short
occasional activations of the other two patterns.
10
Pfeil et al.
Hardware Emulation
11
A)
connections:
excitatory
inhibitory
PN
Glom. i
RN
LN
PN
Glom. i+1
RN
LN
...
PN
B)
LN
Glom. j
RN
C)
output, q=0.0
E)
D)
output, q=1.0
0.8
res. corr
Pfeil et al.
0.6
0.4
0.2
0.0
0.0
0.5
q
1.0
FIGURE 7: (A) Schematic of the insect antennal lobe network. Neuron populations are grouped in glomeruli (outlined by dotted
lines), which exert lateral inhibition onto each other. RNs: receptor neurons (input), PNs: projection neurons (output), LNs:
inhibitory local neurons. Some connections are grayed out to emphasize the connection principle. (B) Correlation matrix of the
input data. (C) Correlation matrix of the output spike rates (PNs) without lateral inhibition, q = 0.0. (D) Correlation of the output
with homogeneous lateral inhibition, q = 1.0. (E) Average pairwise correlation between glomeruli (median 20th (black) and 80th
(gray) percentile) in dependence of the overall strength of lateral inhibition q.
Network Topology
12
Pfeil et al.
We have used 6 RN input streams per glomerulus, projecting in an all-to-all fashion onto 7 PNs, which in turn
projected on 3 LNs per glomerulus.
3.5.2.
Hardware Emulation
(5)
with dPearson (, ) the Pearson correlation coefficient between two vectors. For the input correlation matrix
Cinput , the vector glom.i contained the average firing
rates of the six RNs projecting to the ith glomerulus,
with each element of this vector for one stimulus presentation. For the output correlation matrix Coutput we used
the rates from PNs instead of RNs. Thus, we obtained
10 10 matrices containing the rate correlations for each
pair of input or output channels.
Figure 7B depicts the correlation matrix Cinput for the
input firing rates. When no lateral inhibition is present,
Cinput matches Coutput (Figure 7C). We have systematically varied the strength of lateral inhibition by scaling all
inhibitory weights by a factor q, with q = 0 for zero lateral inhibition and q = 1 for inhibition set to its maximal
strength. With increasing lateral inhibition, off-diagonal
values in Coutput approach zero and output channel correlation is virtually gone (Figure 7D). The amount of
residual correlation to be present in the output can be
controlled by adjusting the strength of lateral inhibition
(Figure 7E).
Taken together, we demonstrated the implementation
of an olfaction-inspired network to remove correlation between input channels on the Spikey chip. This network
can serve as a preprocessing module for data analysis applications to be implemented on the Spikey chip. An
interesting candidate for such an application is a spiking
network for supervised classification, which may benefit strongly from reduced channel correlations for faster
learning and better discrimination (Husler et al., 2011).
3.6.
Network Topology
The LSM consists of two major components: the recurrent liquid network itself and a spike-based classifier.
A general purpose liquid needs to meet the separation
property (Maass et al., 2002), which requires that different inputs are mapped to different outputs, for a wide
range of possible inputs. Therefore, we use a network
topology similar to the one proposed by Bill et al. (2010)
(Figure 8). It consists of an excitatory and inhibitory
population both having recurrent as well as feedforward
connections.
The readout is realized by means of a tempotron (Gtig
& Sompolinsky, 2006), which is compatible with our hardware due to its spike-based nature. Furthermore, its modest single neuron implementation leaves most hardware
resources to the liquid. The afferent synaptic weights are
trained with the method described in Gtig & Sompolinsky (2006), which effectively implements gradient descent
dynamics. Upon training, the tempotron distinguishes
13
Pfeil et al.
0.1
1.5nS
0.2
2nS
Inh
B)
0.8
correctness
0.6
0.1
3nS
0.4
0.2
During learning, weights are updated as follows
Input
Exc
(
0
correct
0.05
(-50, -100)
(-100, -150)
P
wjn =
2nS
(n) ti,j <tmax K(tmax ti,j ) erroneous,
input interval in (ms, ms)
(7)
where wjn is the weight update corresponding to the FIGURE 8: (A) Network architecture of the liquid network with
up to 191 neurons, leaving at least one neuron for the tempotron.
jth afferent neuron after the nth learning iteration with
The ratio of excitatory to inhibitory neurons is 80:20. Each neulearning rate (n). The spike time of the tempotron, ron receives 4 inputs from a pool of 32 excitatory and 32 inor otherwise the time of highest membrane potential, is hibitory sources, respectively. (B) The LSM classification perfordenoted with tmax . In other words, for trials where an mance measured over 200 samples after 1000 training iterations
erroneous spike was elicited, the excitatory afferents with for both hardware (white) and software (grey) implementation.
a causal contribution to this spike are weakened and in- Both implementations classify the input segment from 50 ms to
hibitory ones are strengthened according to Equation 7. 100 ms in the past with about 90% correctness. For inputs lyIn case the tempotron did not spike even though it should ing further behind in time, both implementations drop to chance
have, the weights are modulated the other way round, level.
i.e. excitatory weights are strengthened and inhibitory
ones are weakened. This learning rule has been implemented on hardware with small modifications, due to the 3.6.2. Hardware Emulation
conductance-based nature of the hardware synapses (see
The liquid itself does not impose any strong requirebelow).
ments on the hardware since virtually any network is
The tempotron is a binary classifier, hence any task suitable as long as the separation property is satisfied.
needs to be mapped to a set of binary decisions. Here, we We adapted a network from Bill et al. (2010) (Figure 8)
have chosen a simple binary task to evaluate the perfor- which, in a similar form, had already been implemented
mance of the LSM. The challenge was to distinguish spike on our hardware. However, STP was disabled, because at
train segments in a continuous data stream composed of the time of the experiment it was not possible to exclutwo templates with identical rates. In order to gener- sively enable STP for the liquid without severely affecting
ate the input, we cut two long template spike trains into the performance of the tempotron.
segments of 50 ms each. For any cut we then randomly
The hardware implementation of the tempotron repicked the corresponding spike segment from the one or quired more attention, since only conductance-based
the other template, successively composing new spike se- synapses are available. The dependence of spike efficaquences. Additionally, Gaussian noise with a standard cies on the actual membrane potential was neglected, bedeviation of = 1 ms was added to each spike time. The cause the rest potential was chosen to be close to the
samples were then streamed into the liquid one at a time. firing threshold, with the reversal potentials far away.
Tempotrons were trained to identify the origin of a certain However, the asymmetric distance of excitatory and insegment, within the final 50 ms given the liquid activity hibitory reversal potentials from the sub-threshold regime
as input. Not only did this task allows to determine the needed compensation. This was achieved by scaling all
classification capabilities of the LSM, but it also put the excitatory weights by (Vm Einh )/(Vm Eexc ), where
liquids fading memory to the test: classification becomes Vm corresponds to the mean neuron membrane voltage
increasingly difficult, the further back in time the inter- and Eexc /Einh is the excitatory/inhibitory reversal po-
14
Pfeil et al.
15
Pfeil et al.
plasticity rules are difficult, if not impossible to be emulated on this substrate. Also, the total number of neurons and synapses set a hard upper bound on the size of
networks that can be emulated. However, the next generation of our highly accelerated hardware system will
increase the number of available neurons and synapses
by a factor of 103 , and provide extended configurability
for each of these units (Schemmel et al., 2010).
The main purpose of our hardware system is to provide a flexible platform for highly accelerated emulation of
spiking neuronal networks. Other research groups pursue
different design goals for their hardware systems. Some
focus on dedicated hardware providing specific network
topologies (e.g., Merolla & Boahen, 2006; Chicca et al.,
2007), or comprising few neurons with more complex dynamics (e.g., Chen et al., 2010; Grassia et al., 2011). Others develop hardware systems of comparable configurability, but operate in biological real-time, mostly using offchip communication (Vogelstein et al., 2007; Imam et al.,
2012; Choudhary et al., 2012). Purely digital systems
(Merolla et al., 2011) and field-programmable analog arrays (FPAA; Basu et al., 2010) provide even more flexibility in configuration than our system, but have much
smaller speed-up factors.
With the ultimate goal of brain size emulations, there
exists a clear requirement for increasing the size and complexity of neuromorphic substrates. An accompanying
upscaling of the fitting and calibration procedures presented here appears impractical for such orders of magnitude and can only be done for a small subset of components. Rather, it will be essential to step beyond simulation equivalence as a quality criterion for neuromorphic
computing, and to develop a theoretical framework for
circuits that are robust against, or even exploit the inherent imperfections of the substrate for achieving the
required computational functions.
ACKNOWLEDGMENTS
We would like to thank Dan Husmann, Stefan Philipp,
Bernhard Kaplan, and Moritz Schilling for their essential
contributions to the neuromorphic platform, Johannes
Bill, Jens Kremkow, Anders Lansner, Mikael Lundqvist
and Emre Neftci for assisting with the hardware implementation of the network models, Oliver Breitwieser
for data analysis and Venelin Petkov for characterisation
measurements of the Spikey chip.
The research leading to these results has received funding by the European Union 6th and 7th Framework Programme under grant agreement no. 15879 (FACETS),
no. 269921 (BrainScaleS) and no. 243914 (Brain-i-Nets).
Michael Schmuker has received support from the german
ministry for research and education (BMBF) to Bernstein
Center for Computational Neuroscience Berlin (grant no.
16
Pfeil et al.
Grassia, F., Buhry, L., Lvi, T., Tomas, J., Destexhe, A., &
Saghi, S. (2011). Tunable neuromimetic integrated system for emulating cortical neuron models. Front. Neurosci. 5 (134).
Grossberg, S. (1973). Contour enhancement, short term memory, and constancies in reverberating neural networks. Stud.
Appl. Math. 52 (3), 213257.
Gupta, A., Wang, Y., & Markram, H. (2000). Organizing
principles for a diversity of GABAergic interneurons and
synapses in the neocortex. Science 287, 273278.
Gtig, R., & Sompolinsky, H. (2006). The tempotron: a neuron that learns spike timing-based decisions. Nat. Neurosci. 9 (3), 420428.
Husler, C., Nawrot, M. P., & Schmuker, M. (2011). A spiking
neuron classifier network with a deep architecture inspired
by the olfactory system of the honeybee. In 2011 5th International IEEE/EMBS Conference on Neural Engineering,
pp. 198202. IEEE Press.
Hines, M., Davison, A. P., & Muller, E. (2009). Neuron and
python. Front. Neuroinformatics 3.
Imam, N., Cleland, T. A., Manohar, R., Merolla, P. A.,
Arthur, J. V., Akopyan, F., & Modha, D. S. (2012). Implementation of olfactory bulb glomerular layer computations
in a digital neurosynaptic core. Front. Neurosci. 6 (83).
Indiveri, G., Chicca, E., & Douglas, R. (2006). A VLSI
array of low-power spiking neurons and bistable synapses
with spike-timing dependent plasticity. IEEE Trans. Neural Netw. 17 (1), 211221.
Indiveri, G., Chicca, E., & Douglas, R. (2009). Artificial cognitive systems: From VLSI networks of spiking neurons to
neuromorphic cognition. Cognitive Computation 1 (2), 119
127.
Itti, L., & Koch, C. (2001). Computational modeling of visual
attention. Nat. Rev. Neurosci. 2 (3), 194203.
Jaeger, H. (2001). The echo state approach to analysing
and training recurrent neural networks. Technical Report
GMD Report 148, German National Research Center for
Information Technology.
Kaplan, B., Brderle, D., Schemmel, J., & Meier, K. (2009).
High-conductance states on a neuromorphic hardware system. In Proceedings of the 2009 International Joint Conference on Neural Networks (IJCNN), Atlanta, pp. 15241530.
Kremkow, J., Aertsen, A., & Kumar, A. (2010). Gating of signal propagation in spiking neural networks by balanced and
correlated excitation and inhibition. J. Neurosci. 30 (47),
1576015768.
Kremkow, J., Perrinet, L. U., Masson, G. S., & Aertsen, A.
(2010). Functional consequences of correlated excitatory
and inhibitory conductances in cortical networks. J. Comput. Neurosci. 28 (3), 579594.
17
Pfeil et al.
Merolla, P., Arthur, J., Akopyan, F., Imam, N., Manohar, R.,
& Modha, D. (2011). A digital neurosynaptic core using
embedded crossbar memory with 45pJ per spike in 45nm.
In Proceedings of the 2011 Custom Integrated Circuits Conference (CICC), pp. 14. IEEE Press.
Schemmel, J., Grbl, A., Meier, K., & Mller, E. (2006). Implementing synaptic plasticity in a VLSI spiking neural network model. In Proceedings of the 2006 International Joint
Conference on Neural Networks (IJCNN), Vancouver, pp.
16. IEEE Press.
Neftci, E., & Indiveri, G. (2010). A device mismatch compensation method for VLSI neural networks. In Proceedings of
the 2010 Biomedical Circuits and Systems Conference (BioCAS), pp. 262265. IEEE Press.
Schmuker, M., & Schneider, G. (2007). Processing and classification of chemical data inspired by insect olfaction. Proc.
Natl. Acad. Sci. USA 104 (51), 2028520289.
18
Pfeil et al.
Schrader, S., Diesmann, M., & Morrison, A. (2010). A compositionality machine realized by a hierarchic architecture
of synfire chains. Front. Comput. Neurosci. 4, 154.
Schrader, S., Grn, S., Diesmann, M., & Gerstein, G. (2008).
Detecting synfire chain activity using massively parallel
spike train recording. J. Neurophysiol. 100, 21652176.
Serrano-Gotarredona, R., Oster, M., Lichtsteiner, P., LinaresBarranco, A., Paz-Vicente, R., Gmez-Rodrguez, F.,
Kolle Riis, H., Delbrck, T., Liu, S.-C., Zahnd, S., Whatley, A. M., Douglas, R., Hfliger, P., Jimenez-Moreno, G.,
Civit, A., Serrano-Gotarredona, T., Acosta-Jimnez, A.,
& Linares-Barranco, B. (2006). AER building blocks for
multi-layer multi-chip neuromorphic vision systems. In
Y. Weiss, B. Schlkopf, & J. Platt (Eds.), Adv. Neural. Inf.
Process. Syst., pp. 12171224. Cambridge, MA: MIT Press.
Soltani, A., & Wang, X.-J. (2010). Synaptic computation
underlying probabilistic inference. Nat. Neurosci. 13 (1),
1129.
Song, S., Miller, K. D., & Abbott, L. F. (2000). Competitive
Hebbian learning through spike-timing-dependent synaptic
plasticity. Nat. Neurosci. 3 (9), 919926.
Stopfer, M., Bhagavan, S., Smith, B. H., & Laurent, G. (1997).
Impaired odour discrimination on desynchronization of
odour-encoding neural assemblies. Nature 390 (6655), 704.
Tetzlaff, T., Morrison, A., Timme, M., & Diesmann, M.
(2005). Heterogeneity breaks global synchrony in large networks. In Proceedings of the 30th Gttingen Neurobiology
Conference.
Tsodyks, M. V., & Markram, H. (1997). The neural code
between neocortical pyramidal neurons depends on neurotransmitter release probability. Proc. Natl. Acad. Sci.
USA 94, 719723.
van Vreeswijk, C., & Sompolinsky, H. (1996). Chaos in neuronal networks with balanced excitatory and inhibitory activity. Science 274, 17241726.
Vogels, T. P., & Abbott, L. F. (2005). Signal propagation and
logic gating in networks of integrate-and-fire neurons. J.
Neurosci. 25 (46), 1078610795.
Vogelstein, R., Mallik, U., Vogelstein, J., & Cauwenberghs, G.
(2007). Dynamically reconfigurable silicon array of spiking
neurons with conductance-based synapses. IEEE Trans.
Neural Netw. 18 (1), 253265.
Wilson, R. I., & Laurent, G. (2005). Role of GABAergic inhibition in shaping odor-evoked spatiotemporal patterns in
the drosophila antennal lobe. J. Neurosci. 25 (40), 9069
9079.
Yazdanbakhsh, A., Babadi, B., Rouhani, S., Arabzadeh, E., &
Abbassian, A. (2002). New attractor states for synchronous
activity in synfire chains with excitatory and inhibitory coupling. Biol. Cybern. 86, 367378.
19