Tempotron

From HandWiki
Revision as of 21:13, 14 June 2021 by imported>MainAI5 (linkage)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

The Tempotron is a supervised synaptic learning algorithm which is applied when the information is encoded in spatiotemporal spiking patterns. This is an advancement of the perceptron which does not incorporate a spike timing framework.

It is general consensus that spike timing (STDP) plays a crucial role in the development of synaptic efficacy for many different kinds of neurons [1] Therefore, a large variety of STDP-rules has been developed one of which is the tempotron.

Algorithm

Assuming a leaky integrate-and-fire-model the potential [math]\displaystyle{ V(t) }[/math] of the synapse can be described by

[math]\displaystyle{ V(t)= \sum _i \omega _i\sum _{t_i}K(t-t_i) +V_{rest}, }[/math]

where [math]\displaystyle{ t_i }[/math] denotes the spike time of the i-th afferent synapse with synaptic efficacy [math]\displaystyle{ \omega _i }[/math] and [math]\displaystyle{ V_{rest} }[/math] the resting potential. [math]\displaystyle{ K(t-t_i) }[/math] describes the postsynaptic potential (PSP) elicited by each incoming spike:

[math]\displaystyle{ K(t-t_i) = \begin{cases} V_0[\exp(-(t-t_i)/\tau)-\exp(-(t-t_i)/\tau_s)] & t\geq t_i \\ 0 & t\lt t_i \end{cases} }[/math]

with parameters [math]\displaystyle{ \tau }[/math] and [math]\displaystyle{ \tau_s }[/math] denoting decay time constants of the membrane integration and synaptic currents. The factor [math]\displaystyle{ V_0 }[/math] is used for the normalization of the PSP kernels. When the potential crosses the firing threshold [math]\displaystyle{ V_{th} }[/math] the potential is reset to its resting value by shunting all incoming spikes.

Next, a binary classification of the input patterns is needed([math]\displaystyle{ \circ }[/math] refers to a pattern which should elicit at least one post synaptic action potential and [math]\displaystyle{ \bullet }[/math] refers to a pattern which should have no response accordingly). In the beginning, the neuron does not know which pattern belongs to which classification and has to learn it iteratively, similar to the perceptron . The tempotron learns its tasks by adapting the synaptic efficacy [math]\displaystyle{ \omega _i }[/math]. If a [math]\displaystyle{ \circ }[/math] pattern is presented and the postsynaptic neuron did not spike, all synaptic efficacies are increased by [math]\displaystyle{ \Delta \omega _i }[/math] whereas a [math]\displaystyle{ \bullet }[/math] pattern followed by a postsynaptic response leads to a decrease of the synaptic efficacies by [math]\displaystyle{ \Delta \omega _i }[/math] with [2]

[math]\displaystyle{ \Delta \omega _i=\lambda \sum _{t_i\lt t_{max}}K(t_{max}-t_i). }[/math]

Here [math]\displaystyle{ t_{max} }[/math] denotes the time at which the postsynaptic potential [math]\displaystyle{ V(t) }[/math] reaches its maximal value.

It should be mentioned that the Tempotron is a special case of an older paper which dealt with continuous inputs.[3]

Sources

  1. Caporale, N., & Dan, Y. (2008). Spike timing-dependent plasticity: a Hebbian learning rule. Annu Rev Neurosci, 31, 25-46.
  2. Robert Gütig, Haim Sompolinsky (2006): The tempotron: a neuron that learns spike timing-based decisions, Nature Neuroscience vol. 9, no.3, 420-428
  3. Anthony M. Zador, Barak A. Pearlmutter (1996): "VC dimension of an integrate-and-fire neuron model", Neural Computation vol.8, 611-624