• Keine Ergebnisse gefunden

Practical Course on Pattern Recognition

N/A
N/A
Protected

Academic year: 2022

Aktie "Practical Course on Pattern Recognition"

Copied!
3
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Practical Course on Pattern Recognition

Version: 063006.10 Sommer-Semester 2006

Prof. Dr. Stefan Posch Dipl.Bioinform. Andr´e Gohr (andre.gohr@informatik.uni-halle.de)

Institute of Computer Science University Halle

Series 10

Exercise 10.1(4 points)

Define the back-prop learning rule for a multilayer perceptron that also allows connec- tions (edges between neurons) between non-adjacent layers. But all connections stay feed-forward.

Solution 10.1

Notation: kNi denotes the ith neuron in layer k. Layer k has Mk neurons at all.

wkNi(qNj) denotes the weight of the edge ingoing into neuronkNioutgoing from neuron

qNj if any. The multilayer perceptron consists of Llayers. LayerLis the output-layer.

PkNi denotes the set of ”predecessor” neurons ofkNi that are neurons having an outgo- ing edge being directed towards neuronkNi. DkNidenotes the set of ”direct descendant”

neurons ofkNi.

Further on σ(hkNi) = ykNi denotes the answer of neuron kNi being in activation state hkNi. xkNi(qNj) =yqNj denotes the input of neuron kNi coming from neuron qNj. The activation state hkNi of neuron kNi in layer k is determined by all weighted inputs (weighted answers of ”predeseccor” neurons):

hkNi = P

dNz∈Pk N

i

xkNi(dNz)wkNi(dNz) = P

dNz∈Pk N

i

ydNz wkNi(dNz).

~

yL = (yLN1, . . . , yLN

M L) is the vector of outputs of the perceptron. The error-function is denoted by E(~yL, ~t,w) =

ML

P

i=1

E(yLNi, ti,w). ~t= (t1, . . . , tML) is the vector of target outputs the perceptron should give for a given data-set. wdenotes the set of all weights of the perceptron.

1

(2)

Weights of edges ingoing into any neuronLNi of layerL:

4wLNi(kNj) = −ε ∂E(~y, ~tL,w)

∂wLNi(kNj) with: kNj ∈ PLNi

= −ε

ML

X

r=1

∂E(yLNr, tr,w)

∂wLNi(kNj)

= −ε

ML

X

r=1

∂E(yLNr, tr,w)

∂yLNr

∂yLNr

∂hLNr

∂hLNr

∂wLNi(kNj)

= −ε

ML

X

r=1

∂E(yLNr, tr,w)

∂yLNr σ0(hLNr)

∂ P

dNz∈PLNr

ydNz wLNr(dNz)

∂wLNi(kNj)

P

dNz∈P LNr

ydNzwk N

i(dNz)

∂wLN

i(kNj) is always equal to zero unlessdNz =kNj andLNr =LNi. Hence we get:

4wLNi(kNj) = −ε ∂E(yLNi, ti,w)

∂yLNi σ0(hLNi)ykNj

= −ε δLNiykNj (1)

with: δLNi = ∂E(y∂yLNi,ti,w)

LNi

σ0(hLNi).

Weights of edges ingoing into any neuronL−1Ni of layerL−1:

4wL−1Ni(kNj) = −ε ∂E(~y, ~tL,w)

∂wL−1Ni(kNj) with: kNj ∈ PL−1Ni

= −ε

ML

X

r=1

∂E(yLNr, tr,w)

∂yLNr

∂yLNr

∂hLNr

∂hLNr

∂wL−1Ni(kNj)

= −ε

ML

X

r=1

∂E(yLNr, tr,w)

∂yLNr σ0(hLNr) X

dNz∈PLNr

∂ydNzwLNr(dNz)

∂ydNz

∂ydNz

∂wL−1Ni(kNj)

= −ε

ML

X

r=1

δLNr X

dNz∈PLNr

∂ydNzwLNr(dNz)

∂ydNz

∂ydNz

∂hdNz

∂hdNz wL−1Ni(kNj)

= −ε

ML

X

r=1

δLNr X

dNz∈PLNr

wLNr(dNz0(hdNz) X

sNt∈PdNz

∂ysNtwdNz(sNt) wL−1Ni(kNj)

∂ysNtwdNz(sNt)

wL−1Ni(kNj) is always equal to zero unless sNt = kNj and dNz = L−1Ni. Hence we get:

2

(3)

4wL−1Ni(kNj) = −ε

ML

X

r=1

δLNr wLNr(L−1Ni0(hL−1Ni)ykNj

4wL−1Ni(kNj) = −ε ykNjσ0(hL−1Ni)

ML

X

r=1

δLNr wLNr(L−1Ni)

All outgoing edges of neuron L−1Ni are ingoing into neurons of layer L since layer L is the last layer. One may state this fact more general: All outgoing edges of neuron

L−1Ni are ingoing into one neuron of DL−1Ni.

4wL−1Ni(kNj) = −ε ykNj σ0(hL−1Ni) X

zNv∈DL−1

Ni

δzNvwzNv(L−1Ni)

= −ε ykNj σ0(hL−1NiL−1Ni (2)

with: δL−1Ni = P

zNv∈DL−1Ni

δzNv wzNv(L−1Ni).

Weights of edges ingoing into any neuron mNi of layer m in analogy to the previous derivation (especially equation (2)):

4wmNi(kNj) =−ε ykNjσ0(hmNimNi with: kNj ∈ PmNi (3)

with: δmNi = P

zNv∈DmN

i

δzNvwzNv(mNi).

3

Referenzen

ÄHNLICHE DOKUMENTE

A new approach combining dilatometric experiments and numerical analyses of thermal strains in layered ceramics has been presented, which allows simple way to determine the

Immediately follows the control action, the state satisfies the condition xo=x1=x2, because control action 2 has reset the values of xland x2 to the current disturbance level x o

Field of phrasemes like somatisms in order to language unit character is taken into account in the first section as well as dependence of phrasemes like their meaning explanation

We present an investigation of the structural and magnetic correlations during the remagnetization process of a gradient nanocrystalline multilayer with uniform uniaxial

We derive a cost- sensitive perceptron learning rule for non-separable classes, that can be extended to multi-modal classes (DIPOL) and present a natural cost-sensitive extension of

In an experiment on learning from simulated weak and strong feedback, we confirmed convergence for learning from weak feedback, with faster convergence for stronger feedback and

Works on subsequent academic success often induce strong biases such as the intuition that spelling mistakes indicate cognitive performance. The study of language and context that

This finding is roughly consistent from our finding in this paper for bilateral Hungarian-Slovenian agri-food trade flows where high vertically differentiated products