• Keine Ergebnisse gefunden

Conditionally Positive Semi-Definite Kernels

N/A
N/A
Protected

Academic year: 2021

Aktie "Conditionally Positive Semi-Definite Kernels"

Copied!
2
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Wissenschaftliches Rechnen II/Scientific Computing II

Sommersemester 2016 Prof. Dr. Jochen Garcke Dipl.-Math. Sebastian Mayer

Exercise sheet 8 To be handed in on Thursday, 16.06.2016

Conditionally Positive Semi-Definite Kernels

The main idea of this exercise sheet is to give you a thorough understanding how conditio- nal positive semi-definiteness (cpsd) is naturally related to regularized least-squares pro- blems with penalty functionals J that have non-trivial kernel ker(J) = {f : J (f ) = 0}.

You will mostly consider the univariate cubic smoothing spline on the Sobolev space W

2

:= W

2

([0, 1]) as introduced in Sheet 5, H2. We stress that in this setting cpsd is not needed from a practical point of view since we have nice psd reproducing kernels.

Nevertheless this setting serves as a good demonstration and allows to avoid to much technicalities.

The most important exercise in this regard is H3, which we strongly recommend to work on. The group exercises should give you a profound preparation for H3.

Recall that for given data (x

1

, y

1

), . . . , (x

N

, y

N

) with pairwise different sample points x

i

∈ [0, 1], the cubic smoothing spline is given by ˆ f = argmin

f∈W2

R

`2,reg

(f ). The regularized empirical risk is given by

R

`2,reg

(f) = 1 N

N

X

i=1

(y

i

− f (x

i

))

2

+ λJ

2

(f )

with penalty functional J

2

(f) = kP

1

f k

2W2

, where P

1

denotes the orthogonal projection onto W

02

= {f ∈ W

2

: f (0) = f

0

(0) = 0}.

1 Group exercises

G 1. Provide the proof details for Lemma 51 given in the lecture.

G 2. Let R(x, y) and R

1

(x, y) be the kernels defined in Sheet 5, H2. Imitating the approach given in the lecture notes on p. 35/36, derive the system of linear equations which determines the solution ˆ f . Then show that you can replace R

1

by R in the kernel matrix which appears in the derived linear system. Hint: Choose as the first two elements of your ONB the functions φ

1

(x) = 1, φ

2

(x) = x.

G 3. Choose two arbitrary distinct points t

1

, t

2

from the set of sample points

{x

1

, . . . , x

n

}, say w.l.o.g. t

1

= x

1

, t

2

= x

2

. Let p

1

, p

2

be the unique polynomials of

degree 1 which solve p

i

(t

j

) = δ

ij

for i, j ∈ {1, 2}, where δ

ij

denotes the Kronecker delta

(p

1

, p

2

form a so-called Lagrange basis of Π

1

= span{φ

1

, φ

2

}). As you have proved in G2

(2)

the smoothing spline has the form

f(x) = ˆ α

0

+ α

1

x

| {z }

affine part

+

N

X

j=1

z

j

R

1

(x, x

j

)

| {z }

kernel part

(8)

a) Show that the conditions P

N

i=1

z

i

= P

N

i=1

z

i

x

i

= 0, which you have derived in G2, are equivalent to P

N

i=1

z

i

p

1

(x

i

) = P

N

i=1

z

i

p

2

(x

i

) = 0.

b) Use a) to show that the kernel part is contained in the set P

1

(V

0

), where V

0

:= {f ∈ W

2

: f (t

1

) = f (t

2

) = 0}.

2 Homework

H 1. Let β ∈ N. Show that the thin plate spline (TPS) kernel

k

β

: R

d

× R

d

→ R , (x, y) 7→ (−1)

β+1

kx − yk

2

log(kx − yk

2

) is conditionally positive semi-definite of order β + 1.

(2 Punkte) H 2. The so-called sigmoid kernel

k : R

d

× R

d

→ R , (x, y) 7→ k(x, y) = tanh(κx

T

y + ν),

where κ >, ν < 0 and tanh denotes the tangens hyperbolicus, is often used in the context of artifical neural networks. Show that k is not conditionally positive semi-definite. Hint:

Construct a counter example which exploits the fact that lim

t→±∞

tanh

0

(t) = 0.

(4 Punkte) H 3. (A different perspektive on cpsd)

Consider again the Sobolev space W

2

. The starting point of G2 was to use the orthogonal decomposition

W

2

= Π

1

⊕ W

02

,

which is reflected in the representation (8) of the smoothing spline ˆ f . An alternative representation of ˆ f is of the form

f ˆ = I

(t1,t2)

( ˆ f ) + g,

where I

(t1,t2)

( ˆ f)(x) = p

1

(x) ˆ f (t

1

) +p

2

(x) ˆ f(t

1

) is the linear interpolation of ˆ f in the points t

1

, t

2

, and g ∈ V

0

.

a) Using representation (8), determine I

(t1,t2)

( ˆ f ) and g.

b) Determine the reproducing kernels of the subspaces V

0

and P

1

(V

0

).

(10 Punkte)

2

Referenzen

ÄHNLICHE DOKUMENTE

These emphasize the importance of a clear process and standard for climate impact assessment at multiple startup support selection stages, the need to make investment

In this paper we suggest a hybrid method for computing the smallest eigenvalue of a symmetric and positive definite Toeplitz matrix which takes advantage of two types of

Underline all nouns green which are good for sea turtles.. Underline all nouns red which are bad for

Prove and explain a few main applications of representation theory to finite groups, i.e.. Focus on the main ideas and stress that it took 60 years to find a purely

Appendix A.3], briefly recall the notions of a strict monoidal category and their functors [EGNO15, Section 2.8] or [Kock04, Section 3.2], of a symmetric monoidal category, and

In the case m = 1, the problem to describe all functions preserving positive semi-definiteness and some related problems have a long history and were completely solved by

This work has been digitalized and published in 2013 by Verlag Zeitschrift für Naturforschung in cooperation with the Max Planck Society for the Advancement of Science under

Publisher’s Note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.. The original article can be found online