• Keine Ergebnisse gefunden

3.4 Learning

3.4.3 Reinforcement Learning

ˆ Unpaired actual output spikes that need to be deleted are put into the setD.

ˆ Unpaired desired output spike times are put into the set J, i.e. the set of spikes that have to be inserted.

To clarify, S contains pairs of “paired” actual and desired spike times, D contains the times of all unpaired actual spikes, andJ the times of unpaired desired spike times. With the PSP sum as above, the E-Learning rule is then

∆wi

tins∈J

λi(tins)− ∑

tdel∈D

λi(tdel) +γr τq2

(tact,tdes)∈S

(tact−tdesi(tact)

⎦ . (3.37) γ is the learning rate, and γr is a factor to scale spike shifting relative to deletion and insertion.

The former two terms of the rule correspond to ReSuMe, except the kernel is not a simple exponential decay. The advantage of E-Learning is that the weight changes for spikes close to their desired location are scaled with the distance, which improves convergence and consequentially memory capacity.

3.4.2.3.3 FP-Learning.

FP-Learning [MR ¨OS14] was devised to remedy a central problem in learning rules like ReSuMe and others. Any erroneous or missing spike “distorts” the time course of the membrane potential behind it compared to the desired final state. This creates a wrong environment for the learning rule, and weight changes can potentially be wrong. Therefore, the FP-Learning algorithm stops the learning trial as soon as it encounters any spike output error. Additionally, FP-Learning introduces a margin of tolerable error for the desired output spikes. An actual output spike should be generated in the window of tolerance [td−ϵ, td+ϵ] with the adjustable margin ϵ. Weights are changed on two occasions:

1. If a spike occurs outside the window of tolerance for anytdat timeterr, then weights are depressed by ∆wi ∝ −λi(terr). This also applies if the spike in question is the second one within a given tolerance window.

2. Ift=td+εand no spike has occured in the window of tolerance, thenterr=td+ε and ∆wi ∝λi(terr).

In both cases, the learning trial immediately ends, to prevent that the “distorted” mem-brane potential leads to spurious weight changes. Because of this property, this rule is also referred to as “First Error Learning”.

is higher, it will keep the change and otherwise discard it.

In this way, e.g. irregularly spiking neural networks can estimate the gradient of the reward signal and perform a stochastic gradient ascent on the expected reward [XS04]. In their model, Xie and Seung introduce stochastic spiking neurons i, which receive input from input neurons j. The strength of the connection between j and i is given by the synaptic weight wij. The input current into neuron ithen given by

Ii(t) =∑

j

wijhij(t) (3.38)

wherehij evolves according to τsdhij

dt +hij =∑

a

δ(t−Tjaija (3.39) where ζija is a binary random variable modelling the stochastic nature of synaptic trans-mission andTja is the time of the ath spike in input neuron j. The input current Ii(t) is converted to an instantaneous firing rateλi(t) by

λi(t) =fi(Ii(t)) (3.40)

wherefi is the current-discharge relationship.

The learning rule is then given by

∆wij =ηReij (3.41)

whereR is a reward signal,η is a learning rate andeij is an eligibility trace given by eij =

T 0

dtΦi(Ii) [si(t)−fi(Ii)]hij (3.42) whereT is the length of the learning episode and si(t) =∑

aδ(t−Tia) is the spike train in neuron i. Φi(Ii) is a function that scales the weight changes depending on the current firing rate. Learning works, because if the actual activation of the neuron is larger than the instantanous firing rate and this is rewarded (R > 0), then weights from positively contributing input neurons are increased to increase the instantaneous firing rate upon repetition. Similarly, if the actual output of the neuron is above the instantaneous firing rate and this is punished (R < 0), weights are changed to decrease the firing rate. The same mechanism also applies for an actual activation below the firing rate, such that the firing rate is changed in the desired direction. By this learning mechanism, a gradient ascent on the reward is performed.

For constant reward and an output spike train that is enforced by a teacher, this is very similar to theδ-rule.

In a study by Fiete et al., reinforcement learning has been shown to be applicable to more realistic neuron models, where the exploration is done by a perturbation of the conductance of the neuron [FS06].

3.4.3.1 Learning in Recurrent Networks

So far, only feed-forward classifiers on more or less complex patterns were discussed. In this section, I want to discuss learning in recurrent networks. This setup if of particular interest, because in biological neuronal networks, at least some degree of feedback input can be expected.

3.4.3.1.1 Hopfield networks

Hopfield networks [Hop82, Hop07] are arguably the most simple setup for a recurrent network: A population of N rate neurons is interconnected by weights wij, which define the recurrent input. The activation of each neuron i is binary with Si ∈ {−1,1}, where Si = 1 is an activated state and Si = −1 is a silent state. In this setup, the input into each neuron is given by

hi=∑

j

wijSj (3.43)

The activation of the neuron will then be defined by an activation function, which is chosen to be the sign function, such that

Si := sgn

j

wijSj

⎠ (3.44)

Activations of neurons can either all be updated the same time in a synchronous update or one after the other in an asynchronous update.

Imprinting P patterns ξµ onto this network can generate stable patterns, that are self-consistent in that they generate input into each neuron that is compatible with its own state. Patterns can even be completed, if a noisy version of the pattern is presented to the network, because an attractor around the original pattern is formed.

In the Hopfield model, a generalized Hebb rule is employed for learning:

wij = 1 N

P

µ=1

ξiµξµj (3.45)

With this learning rule, a number of patterns can be stored in such a neuronal network, such that these patterns form attractors. This is a very simple form of content-addressable memory and associative learning.

3.4.3.1.2 Temporal sequences of patterns

So far, stationary patterns were discussed. However, in biological neuronal networks, sequences of activation patterns are essential. In modelling studies, these activation se-quences are usually considered in closed loop situations, where the last activation pattern in the sequence restarts the sequence (limit cycles). This process can be seen as a stable trajectory in a phase space, where the vector of the states of the neurons is the state of the system. This state vector then follows a closed trajectory.

Recurrent neuronal networks, which are able to create continuous, self-sustained patterns of activity, are highly sensitive to noise. Hence, any learning algorithm that attempts to teach elongated activity sequences to recurrent networks needs to take this sensitivity into

account.

Laje and Buonomano introduce a model that produces robust patterns in recurrent neural networks in [LB13]. Their model is based on firing-rate units and generates locally stable trajectories in the phase space.

They use innate trajectories through the phase space as a starting point, that is, sequences of activation that the network generates due to the initialization. During a training period, they present the network with noisy input and apply a learning rule similar to the delta rule to achieve an output equivalent to the original trajectory. Thus, they generate areas of attraction around the desired trajectory, from which the network evolves back onto the trajectory. Thus, noise during learning is beneficial for the stability of learned trajectories.

To derive meaningful output from this recurrent network, they train an output neuron to respond to the learned pattern in a specific way.

Brea et al. [BSP13] introduce another model based on spike-response model neurons with stochastic spiking. They use a set of visible neurons that are part of the training pattern and a set of hidden neurons that can spike at liberty. Hidden neurons enable networks to solve more complex problems, that are not solvable with visible neurons only. In their study, they introduce a membrane potential

ui(t) =u0+

N

j

wijxϵj(t) +xκi(t) (3.46) wherewij is the synaptic strength from neuronjto neuroni,xαk(t) =∑

s=1α(s)xk(t−s) represents the convolution of spike train xk with kernel α and u0 is the resting poten-tial. The postsynaptic kernel is given byϵ(s) = τ 1

1−τ2(exp (−s/τ1)−exp (−s/τ2)) and the adaptation kernel byκ(s) =cexp(−s/τr) for s≥0. Both kernels are zero fors <0.

The spiking process is modelled as a stochastic process based on the deterministic mem-brane potential. The spiking probability of neuroniin time bin tis given by

P(xi(t) = 1|ui(t)) =ρ(ui(t)) (3.47) withρ(u) = 1+exp(−βu)1 .

In this setup, they develop a learning rule from the goal that the distribution of output spike trains in the visible neuronsPw(v) is as similar as possible to the target distribution P(v). To that end they calculate a learning rule that performs a gradient descent on an upper bound on the Kullback-Leibler divergence given by

D(P(v)||Pw(v)) =⟨logP(v)

Pw(v)⟩P(v) (3.48)

The derived learning rule is given by

∆wbatchij

T

i=1

gi(t) (xi(t)−ρi(t))xϵj(t)

{1 ifivisible

logRw(v|h)−¯r ifihidden (3.49) where η is the learning rate. gi(t) = ρ

i(t)

ρi(t)(1−ρi(t)) with ρi(t) = dρ(t)du |u=ui(t), which implies gi(t) =βforρ(U) as defined above. Rw(v|h) is the probability of a visible activity pattern, given the past hidden pattern and ¯ris a constant. During the learning process, spike trains for the visible neurons are sampled from the target distribution v P(v) and imposed on

the visible neurons. Hidden neurons follow the dynamics of the network. The main part of the learning rule is essentially equivalent to the delta learning rule: The difference between the actual and the real activation of each target neuron is correlated with the presynaptic activation.

With this learning rule, a network of spiking neurons can learn to approximate a desired output spike pattern distribution.