|Welcome to The Neuromorphic Engineer|
Biological Models » Neural
Using neuron dynamics for realistic synaptic learning
PDF version | Permalink
Computational tasks such as object and pattern recognition rely on deterministic learning in the brain carried out mostly at the synapses, which link the brain's neurons and shape the overall computational function of a group of neurons.1 Modelers build mathematical abstractions of synaptic learning by approximating the behavior of biological synapses as they try to copy relevant processing functions.2–4 Neuromorphic integrated circuits (ICs) implement transistor-based versions of these mathematical abstractions in order to realize adaptive, error-tolerant emulations of cognitive functions.5,6
In recent years, there has been a steady increase in the size of neuromorphic systems in order to handle more advanced cognitive tasks. This calls for area-efficient circuit implementations, especially of synapses, since biology-derived topologies use far more synapses than neurons,7 which makes synapse size the determining factor in overall IC complexity.6,8,9 While reducing synapse size, ideally the biological accuracy of the synapses' learning function should increase to keep pace with the biologists' and modelers' continuously refined understanding of cognitive functions.10 To date, these conflicting demands on the learning circuit's implementation have not received much attention. In particular, the usual two-step approach of deriving a mathematical model and subsequently building circuits for it tends to yield very complex circuits.5,8 However, co-developing both the mathematical model and circuit implementation could balance both objectives, resulting in a circuit-optimized mathematical model that also exhibits good biological accuracy.
A synapse composes its learning function from the neurons' local-state variables.1But most models of synaptic learning introduce synthetic dynamical variables driven by higher order information such as spike timings.2,3,11 From a hardware perspective, implementing dynamical behavior—especially configurable time constants—consumes area. As a result, using local waveforms in the synaptic learning circuit would both closely mimic the biology and save area-expensive circuitry.
Our learning rule, called local-correlation plasticity (LCP), takes this approach, measuring the relationship between the activities of the triggering (presynaptic) and the receiving (postsynaptic) neurons of a synapse.12The activity of the postsynaptic neuron is naturally included in its membrane voltage u(t), which also has great influence on learning.13,14 The presynaptic neuron's activity is visible to the synapse by a conductance change g(t) in the receiving neuron. Multiplying both variables results in the synaptic strength w121 illustrates the operation of the LCP rule. While u(t) determines the direction of the weight change, g(t) corresponds to a gating amplifier. Synaptic learning is only possible with presynaptic activity.
With our LCP learning rule, the number of dynamical variables can be minimized in the system design. The membrane voltage u(t) is generated by the neuron circuit and fed back to all connected synapses. Similarly, the waveform of the synaptic input g(t) needs to be generated only once per sending neuron. Thus, all waveform-generation circuits are separated from the individual synapse. This allows for a compact synapse circuit that essentially consists of a differential pair carrying out the difference computation and multiplication of equation (1).15 The waveform-sharing by the synapses fits nicely into the matrix structure used in neuromorphic chip design6,8 (see Figure 2). Each synapse row connects to one postsynaptic neuron. A synapse column is driven by two input waveforms from which each synapse chooses one waveform.9This architecture provides sufficient flexibility for mapping typical network models9 while maximally reusing waveform-generation circuitry, thus making for an area-saving design. However, we wondered whether this approach could account for modern learning paradigms.
One of the most influential models for synaptic learning, spike-timing-dependent plasticity (STDP),11 accounts for the synaptic weight increase and decrease observed when applying a single spike of the pre- and postsynaptic neuron at fixed time intervals to the synapse.16 In the LCP rule, the timing dependencies arise naturally from the correlation between postsynaptic voltage and presynaptic conductance change, that is, the time windows are inherent in u(t) and g(t) (compare Figure 1). Figure 3 shows that a circuit implementation of the LCP rule,15 despite not being conceived specifically for STDP reproduction, can approximate the typical STDP window. It is also evident from Figure 3 that even, for Three Sigma manufacturing deviations, the STDP window only changes in its scaling, not its typical shape. This shows that our complexity-reducing approach results in a very robust learning implementation.
Recent computational analysis indicates that learning at the synapse is significantly more involved than a simple STDP dependency, so a generic third variable needed to be postulated.10 Experimental evidence also supports this postulate, with learning dependent on more complex spike orders,2 on spike rate,14 or on alterations of the postsynaptic membrane voltage.13,14 Figure 4 shows that the circuit realization of the LCP rule can replicate such a third variable. In this example, LCP replicates a pulse-triplet dependency with a high degree of biological veracity,2 even though the single synapses are much simpler than typical STDP implementations.5,8
The synaptic learning expressed in our LCP rule combines high robustness and area efficiency in circuit terms with a biology-driven mathematical formulation. The LCP rule and its circuit realization can reproduce a wide range of complex biological learning phenomena4,15 on a par with state-of-the-art, biology-oriented learning rules.4,17 The circuit realization of the LCP rule is thus well suited to building large-scale neuromorphic systems with a much higher degree of biological accuracy than current approaches. We are involved with several neuromorphic hardware projects that provide an ideal substrate for applying the LCP rule to adaptive closed-loop interfaces to biological neural tissue or to biology-derived, very large scale integration information processing. Our future work will continue the co-development approach, especially modulation of the learning function (metaplasticity) of the LCP rule4 in the circuit realization.
Tell us what to cover!
If you'd like to write an article or know of someone else who is doing relevant and interesting stuff, let us know. E-mail the editor and suggest the subject for the article and, if you're suggesting someone else's work, tell us their name, affiliation, and e-mail.