• Keine Ergebnisse gefunden

3.1.2 Constrained Tikhonov Regularization

In some cases, one knows in advance that that the true solutionπ‘₯†lies in some convex subset 𝐢 βŠ‚ 𝑋. In those cases, the result can often be improved by replacing the Tikhonov functional𝐽uοΏ½in (3.7) with

𝐽uοΏ½uοΏ½(π‘₯) ∢= βˆ₯𝑇π‘₯ βˆ’ 𝑦βˆ₯2+ 𝛼βˆ₯π‘₯βˆ₯2+ πœ’uοΏ½(π‘₯), where

πœ’uοΏ½(π‘₯) =⎧{

⎨{⎩0, π‘₯ ∈ 𝐢,

∞, otherwise is the indicator function of the set𝐢. Thus, one solves

𝑅uοΏ½,u�𝑦uοΏ½ = argmin

u�∈uοΏ½ 𝐽uοΏ½(π‘₯).

If𝐢is closed and convex, a minimizer of𝐽uοΏ½uοΏ½exists and is uniquely determined (see for example [EHN96, Theorem 5.15]). Since𝑅uοΏ½,uοΏ½is non-linear, spectral theory can not be employed in the convergence analysis. Convergence of orderβˆšπ›Ώfor this method was shown in [Neu88] under theprojected source condition

π‘₯†= 𝑃uοΏ½(π‘‡βˆ—πœ”) for someπœ” ∈ π‘Œ, (3.9) where𝑃u�∢ 𝑋 β†’ 𝑋is the metric projection onto𝐢. This condition is similar to a HΓΆlder source conditionπ‘₯† βˆˆβ„›((π‘‡βˆ—π‘‡)1/2) =β„›(π‘‡βˆ—)with exponent Β½, but can be significantly weaker depending on the constraint set𝐢.

3.2 Discretization of Linear Inverse Problems

In practice, the problem (3.1) can only be solved in finite dimensions. Therefore, the continuous formulation has to be discretized. This is done by introducing orthogonal projectionsπ‘ƒβ„Ž βˆˆβ„’(𝑋)andπ‘„β„Ž βˆˆβ„’(π‘Œ)with finite dimensional range. Here,β„Ž > 0is a discretization parameter. Then (3.1) is replaced by

π‘‡β„Žπ‘₯ ∢= π‘„β„Žπ‘‡π‘ƒβ„Žπ‘₯ = π‘„β„Žπ‘¦.

This equation is then solved forπ‘₯ ∈ β„›(π‘ƒβ„Ž)by the methods introduced in the previ-ous sections. The aim of this section is to investigate the convergence of discretized Tikhonov regularization under a convex constraint.

In the following, it is assumed thatβˆ₯π‘‡β„Žβˆ’ 𝑇βˆ₯ β†’ 0asβ„Ž β†’ 0. To check this property in concrete cases, the following lemma may be useful.

Lemma 3.6. The following statements are equivalent:

1. βˆ₯π‘‡β„Žβˆ’ 𝑇βˆ₯ β†’ 0asβ„Ž β†’ 0

2. βˆ₯𝑇(1 βˆ’ π‘ƒβ„Ž)βˆ₯ β†’ 0andβˆ₯(1 βˆ’ π‘„β„Ž)𝑇βˆ₯ β†’ 0

3. 𝑇is compact,π‘ƒβ„Ž β†’ 1pointwise onβ„›(π‘‡βˆ—)andπ‘„β„Žβ†’ 1pointwise onβ„›(𝑇) Proof. (1.) implies (2.) since

βˆ₯(1 βˆ’ π‘„β„Ž)𝑇βˆ₯ ≀ βˆ₯π‘‡β„Žβˆ’ π‘„β„Žπ‘‡βˆ₯ + βˆ₯π‘‡β„Žβˆ’ 𝑇βˆ₯ ≀ 2βˆ₯π‘‡β„Žβˆ’ 𝑇βˆ₯, and similarly forβˆ₯𝑇(1 βˆ’ π‘ƒβ„Ž)βˆ₯. The reverse implication holds due to

βˆ₯π‘‡β„Žβˆ’ 𝑇βˆ₯ ≀ βˆ₯(1 βˆ’ π‘„β„Ž)π‘‡π‘ƒβ„Žβˆ₯ + βˆ₯𝑇(1 βˆ’ π‘ƒβ„Ž)βˆ₯.

For the rest, note that due toβˆ₯𝑇(1βˆ’π‘ƒβ„Ž)βˆ₯ = βˆ₯(1βˆ’π‘ƒβ„Ž)π‘‡βˆ—βˆ₯, it suffices to show the statements forπ‘„β„Ž. The implication (2.) β‡’ (3.) holds since𝑇 is the limit of finite-dimensional operatorsπ‘„β„Žπ‘‡ and hence compact, and sinceπ‘„β„Žπ‘¦ β†’ 𝑦for all𝑦 = 𝑇π‘₯withπ‘₯ ∈ 𝑋. To see the reverse, assume thatβˆ₯(1 βˆ’ π‘„β„Ž)𝑇βˆ₯ ↛ 0. Then there are𝑐 > 0and sequences (β„ŽuοΏ½)uοΏ½βˆˆβ„• βŠ‚ ℝand(π‘₯uοΏ½)uοΏ½βˆˆβ„• βŠ‚ 𝑋withβ„ŽuοΏ½ β†’ 0,βˆ₯π‘₯uοΏ½βˆ₯ = 1and

βˆ₯(1 βˆ’ π‘„β„ŽuοΏ½)𝑇π‘₯uοΏ½βˆ₯ β‰₯ 𝑐.

By compactness, we can assume without loss of generality that 𝑇π‘₯uοΏ½ β†’ 𝑦 ∈ β„›(𝑇). Then the previous inequality implies that(1 βˆ’ π‘„β„Ž)𝑦 ↛ 0. On the other hand, uniform boundedness of (π‘„β„Ž)β„Ž>0 implies that π‘„β„Ž β†’ 1 pointwise also on β„›(𝑇), which is a contradiction.

We cite the following convergence result for the discretized solution under HΓΆlder source conditions in the unconstrained case from [PV90].

Theorem 3.7. Let π‘ƒβ„Ž,π‘„β„Ž and π‘‡β„Ž be as above, 𝑦 ∈ β„›(𝑇) and 𝑦uοΏ½ ∈ π‘Œ with βˆ₯𝑦 βˆ’ 𝑦uοΏ½βˆ₯ ≀ 𝛿. Assume thatβˆ₯𝑇 βˆ’ π‘‡β„Žβˆ₯ β†’ 0asβ„Ž β†’ 0, and that the regularization parameter𝛼 > 0is chosen such that

𝛼 β†’ 0, 𝛿𝛼 β†’ 0,2 βˆ₯𝑇(1 βˆ’ π‘ƒβ„Ž)βˆ₯2

𝛼 ≀ 𝑐 (3.10)

as𝛿, β„Ž β†’ 0for some𝑐 > 0. Denote by Μ„π‘₯ βˆˆβ„›(π‘ƒβ„Ž)the discrete Tikhonov-regularized solution fulfilling

(π‘‡β„Žβˆ—π‘‡β„Ž+ 𝛼) Μ„π‘₯ = π‘‡β„Žβˆ—π‘¦uοΏ½.

Then uοΏ½,β„Žβ†’0lim Μ„π‘₯ = π‘₯†.

If, in addition,

π‘₯† βˆˆβ„›((π‘‡βˆ—π‘‡)uοΏ½)

3.2 Discretization of Linear Inverse Problems for some0 < πœ‡ ≀ 12, and

𝑐1𝛼 ≀ 𝛿2uοΏ½+12 + βˆ₯𝑇(1 βˆ’ π‘ƒβ„Ž)βˆ₯2 ≀ 𝑐2𝛼 for some𝑐1, 𝑐2> 0, then

βˆ₯ Μ„π‘₯ βˆ’ π‘₯†βˆ₯ =uοΏ½(𝛿2uοΏ½+12uοΏ½ + βˆ₯𝑇 βˆ’ π‘‡β„Žβˆ₯2uοΏ½) (3.11) as𝛿, β„Ž β†’ 0.

Unfortunately, the proof in [PV90] relies on spectral theory and therefore does not generalize to the constrained setting.

To address this problem, spectral theory can be replaced by variational techniques in regularization theory developed in the last decade. They were first introduced in [HKPS07] for problems with non-linear, non-smooth operators, but also turned out useful to handle more general data fidelities and penalties as well as constrained problems. In the following, we will prove a convergence result similar to Theorem 3.7 using these techniques.

Even though we are only interested in the Tikhonov case, the convergence theorem will be presented in a more abstract way, for which some additional notation is needed.

Assume that we are given an ideal data fidelity functional uοΏ½ ≑ uοΏ½uοΏ½β€ βˆΆ π‘Œ β†’ [0, ∞]

with

uοΏ½(𝑦) = 0 ⟺ 𝑦 = 𝑦†,

and anempirical data fidelity functionaluοΏ½ ≑uοΏ½uοΏ½u�∢ π‘Œ β†’ ℝ βˆͺ {∞}, both convex, which are connected by the error assumption

𝐴1 uοΏ½(𝑦) βˆ’ πœ€uοΏ½ ≀uοΏ½(𝑦) βˆ’uοΏ½(𝑦†) ≀ 𝐴(uοΏ½(𝑦) + πœ€uοΏ½) βˆ€π‘¦ ∈ π‘Œ

for some 𝐴 β‰₯ 1and πœ€uοΏ½ > 0 such that πœ€uοΏ½ β†’ 0 as𝛿 β†’ 0. The convexregularization penaltyis denoted byβ„›βˆΆ 𝑋 β†’ ℝ βˆͺ {∞}, and the discretization error is controlled by the assumption

𝐡1 uοΏ½(𝑇π‘₯) βˆ’ πœ‚β„Žβ„›(π‘₯) ≀uοΏ½(π‘‡β„Žπ‘₯) ≀ 𝐡(uοΏ½(𝑇π‘₯) + πœ‚β„Žβ„›(π‘₯)) βˆ€π‘₯ ∈ 𝐢 βŠ‚ 𝑋 (3.12a) for𝐡 β‰₯ 1andπœ‚β„Ž> 0withπœ‚β„Žβ†’ 0asβ„Ž β†’ 0, and where𝐢 βŠ‚ 𝑋is the convex constraint set. Then (generalized) discrete constrained Tikhonov regularization consists in solving the optimization problem

Μ„π‘₯ ∈ argmin

u�∈uοΏ½ (uοΏ½(π‘‡β„Žπ‘₯) + 𝛼ℛ(π‘₯)). (3.13)

In the variational formalism, convergence is usually investigated in terms of the Breg-man distanceof the convex regularization penalty. It is defined as

uοΏ½β„›(π‘₯, π‘₯†) ≑uοΏ½u�ℛ†(π‘₯, π‘₯†) =β„›(π‘₯) βˆ’β„›(π‘₯†) βˆ’ βŸ¨πœ‰β€ , π‘₯ βˆ’ π‘₯β€ βŸ©,

whereπœ‰β€  ∈ πœ•β„›(π‘₯†)is a fixed element in the subgradient of β„› at π‘₯†. Note that in general,uοΏ½β„› is neither symmetric nor does it satisfy a triangle inequality.

An important part in the variational approach consists in replacing the spectral source condition (3.5) by a condition in the form of a variational inequality.

Definition 3.8. π‘₯β€ βˆˆ 𝑋is said to satisfy an (additive)variational smoothness assumption if there are𝛽 > 0and a concave index functionπœ“βˆΆ ℝ+ β†’ ℝwith

𝛽uοΏ½β„›(π‘₯, π‘₯†) ≀ℛ(π‘₯) βˆ’β„›(π‘₯†) + πœ“(uοΏ½(𝑇π‘₯)) βˆ€π‘₯ ∈ 𝐢. (3.14) Quadratic Tikhonov regularization is the special case

uοΏ½(𝑦) = 12βˆ₯𝑦 βˆ’ 𝑦†βˆ₯2, uοΏ½(𝑦) = 12βˆ₯𝑦 βˆ’ 𝑦uοΏ½βˆ₯2, and

β„›(π‘₯) = 12βˆ₯π‘₯βˆ₯2+ πœ’uοΏ½(π‘₯).

(3.15)

In this case, assuming thatπ‘₯† ∈ 𝐢, the Bregman distance atπ‘₯ ∈ 𝐢is simply uοΏ½β„›(π‘₯, π‘₯†) = 12βˆ₯π‘₯ βˆ’ π‘₯†βˆ₯2+ πœ’uοΏ½(π‘₯).

To show that (3.12) is fulfilled case with𝐴 = 2,𝐡 = 1and πœ€uοΏ½ = βˆ₯𝑦uοΏ½βˆ’ 𝑦†βˆ₯2 and πœ‚β„Ž= βˆ₯𝑇 βˆ’ π‘‡β„Žβˆ₯2, we use the inequalityβˆ₯π‘Ž + 𝑏βˆ₯2≀ 2βˆ₯π‘Žβˆ₯2+ 2βˆ₯𝑏βˆ₯2to obtain

12uοΏ½(𝑦) βˆ’ πœ€uοΏ½ ≀ 12βˆ₯𝑦 βˆ’ 𝑦uοΏ½βˆ₯2βˆ’ 12βˆ₯π‘¦β€ βˆ’ 𝑦uοΏ½βˆ₯2=uοΏ½(𝑦) βˆ’uοΏ½(𝑦†) and

uοΏ½(𝑦) βˆ’uοΏ½(𝑦†) ≀ βˆ₯𝑦 βˆ’ 𝑦†βˆ₯2+ 12βˆ₯𝑦uοΏ½βˆ’ 𝑦†βˆ₯2≀ 2(uοΏ½(𝑦) + πœ€uοΏ½).

The inequality (3.12a) follows from

uοΏ½(π‘‡β„Žπ‘₯) = 12βˆ₯π‘‡β„Žπ‘₯ βˆ’ 𝑦†βˆ₯2

≀ βˆ₯𝑇π‘₯ βˆ’ 𝑦†βˆ₯2+ βˆ₯(𝑇 βˆ’ π‘‡β„Ž)π‘₯βˆ₯2

≀ βˆ₯𝑇π‘₯ βˆ’ 𝑦†βˆ₯2+ βˆ₯𝑇 βˆ’ π‘‡β„Žβˆ₯2βˆ₯π‘₯βˆ₯2

=uοΏ½(𝑇π‘₯) + πœ‚β„Žβ„›(π‘₯).

3.2 Discretization of Linear Inverse Problems The inequalityuοΏ½(𝑇π‘₯) ≀uοΏ½(π‘‡β„Žπ‘₯) + πœ‚β„Žβ„›(π‘₯)is obtained in the same way by exchanging 𝑇andπ‘‡β„Ž.

Several results on the relations between spectral source conditions and variational smoothness assumptions have been obtained (an overview is given in [Fle13]). The following is from [HY10].

Theorem 3.9. In the quadratic Tikhonov case(3.15)and for a linear operator𝑇 ∢ 𝑋 β†’ π‘Œ, the HΓΆlder source condition(3.5)withπœ‘ = πœ‘uοΏ½ given by(3.6)for0 < πœ‡ ≀ 12 implies(3.14)with πœ“ ∼ πœ‘uοΏ½for𝜈 = 2uοΏ½+12uοΏ½ and𝛽 ∈ (0, 1)arbitrary.

For larger HΓΆlder exponents, this theorem can not hold. In fact, in [HY10] it is also proved that (3.14) withπœ“ ∼ πœ‘uοΏ½ for𝜈 > 1/2impliesπ‘₯† = 0. So in this sense, variational smoothness assumptions only cover HΓΆlder source conditions withπœ‡ ≀ 1/2.

Moreover, in [FH11], equivalence between the projected source condition (3.9) and (3.14) withπœ“(𝑑) ∼ βˆšπ‘‘was shown. This serves as the main motivation to approach the discrete, constrained problem with these methods.

To prove the convergence theorem, some basic results from convex analysis are re-quired:

Definition 3.10. Given a convex, lower semi-continuous function𝑓 ∢ ℝuοΏ½ β†’ ℝ βˆͺ {∞}not constantly∞, itsconjugate functionπ‘“βˆ— is defined by

π‘“βˆ—(𝑦) = sup

uοΏ½βˆˆβ„uοΏ½(𝑦uοΏ½π‘₯ βˆ’ 𝑓 (π‘₯)), 𝑠 ∈ ℝ.

If𝑓 is only defined on a subset ofℝuοΏ½, it is understood to be extended to all of it by∞. Lemma 3.11. The following statements hold:

β€’ 𝑓 (π‘₯) + π‘“βˆ—(𝑦) β‰₯ 𝑦uοΏ½π‘₯,

β€’ 𝑓 (π‘₯) + π‘“βˆ—(𝑦) = 𝑦uοΏ½π‘₯ ⟺ 𝑦 ∈ πœ•π‘“ (π‘₯) ⟺ π‘₯ ∈ πœ•π‘“βˆ—(𝑦), and

β€’ π‘“βˆ—βˆ— ∢= (π‘“βˆ—)βˆ— = 𝑓.

Proof. More general versions of these statements are shown for example in [ET76, Chapter 1].

Theorem 3.12. Let the relations(3.12) hold and Μ„π‘₯be defined as the assumed to be unique minimizer in(3.13). Assume further thatπ‘₯†is the uniqueβ„›-minimizing solution to𝑇π‘₯ = 𝑦†.

1. LetuοΏ½ and β„› be weakly lower semi-continuous and the sets {π‘₯ ∈ 𝑋 ∢ β„›(π‘₯) ≀ 𝑐} be weakly sequentially compact for all𝑐 > 0.1 If𝛼is chosen such that𝛼 β†’ 0and

𝛼(1 uοΏ½(π‘‡β„Žπ‘₯†) βˆ’ infu�∈uοΏ½uοΏ½(π‘‡β„Žπ‘₯)) β†’ 0 (3.16) as𝛿, β„Ž β†’ 0, then

uοΏ½,β„Žβ†’0limuοΏ½β„›( Μ„π‘₯, π‘₯†) = 0.

2. If the variational smoothness assumption(3.14)holds true, then 𝛽uοΏ½β„›( Μ„π‘₯, π‘₯†) ≀ π΄π΅πœ“(𝒆𝒓𝒓)

with 𝒆𝒓𝒓 = (𝐴𝐡 + π΄βˆ’1)β„›(π‘₯†)πœ‚β„Ž+ (1 + 𝐴)πœ€uοΏ½, if𝛼is chosen suitably (see(3.19)below).

Proof. By definition,

uοΏ½(π‘‡β„Ž Μ„π‘₯) + 𝛼ℛ( Μ„π‘₯) ≀uοΏ½(π‘‡β„Žπ‘₯†) + 𝛼ℛ(π‘₯†) (3.17) and thus, using (3.16),

lim sup

uοΏ½,β„Žβ†’0 β„›( Μ„π‘₯) ≀ℛ(π‘₯†). (3.18)

We now show that each sequence of minimizers( Μ„π‘₯uοΏ½)uοΏ½βˆˆβ„• fulfilling the requirements with corresponding errorsπœ€uοΏ½, πœ‚uοΏ½ β†’ 0has a subsequence that converges weakly toπ‘₯†. This then implies that Μ„π‘₯ ⇀ π‘₯†as𝛿, β„Ž β†’ 0.

Due to (3.18) and the compactness assumption onβ„›,( Μ„π‘₯uοΏ½)has a weakly convergent subsequence uοΏ½(uοΏ½)Μ„π‘₯ ⇀ Μƒπ‘₯ as𝑙 β†’ ∞. Then𝑇 Μ„π‘₯uοΏ½(uοΏ½) ⇀ ̃𝑇π‘₯, since every continuous linear operator is weakly continuous. Using the weak lower semi-continuity ofuοΏ½, (3.17) and the error relations (3.12), we obtain

uοΏ½(𝑇 Μƒπ‘₯) ≀ lim infuοΏ½β†’βˆž uοΏ½(𝑇 Μ„π‘₯uοΏ½(uοΏ½))

≀ lim infuοΏ½β†’βˆž 𝐡(uοΏ½(𝑇uοΏ½(uοΏ½) uοΏ½(uοΏ½)Μ„π‘₯ ) + πœ‚β„Žβ„›(π‘₯))

≀ lim infuοΏ½β†’βˆž 𝐡uοΏ½(𝑇uοΏ½(uοΏ½) uοΏ½(uοΏ½)Μ„π‘₯ )

≀ lim infuοΏ½β†’βˆž 𝐴𝐡(uοΏ½uοΏ½(uοΏ½)(𝑇uοΏ½(uοΏ½) uοΏ½(uοΏ½)Μ„π‘₯ ) βˆ’uοΏ½uοΏ½(uοΏ½)(𝑦†))

≀ lim infuοΏ½β†’βˆž 𝐴𝐡(uοΏ½uοΏ½(uοΏ½)(𝑇uοΏ½(uοΏ½)π‘₯†) βˆ’uοΏ½uοΏ½(uοΏ½)(𝑦†) + 𝛼uοΏ½(uοΏ½)(β„›(π‘₯†) βˆ’β„›( Μ„π‘₯uοΏ½(uοΏ½))))

= lim infuοΏ½β†’βˆž 𝐴𝐡(uοΏ½uοΏ½(uοΏ½)(𝑇uοΏ½(uοΏ½)π‘₯†) βˆ’uοΏ½uοΏ½(uοΏ½)(𝑦†))

≀ lim infuοΏ½β†’βˆž 𝐴2𝐡uοΏ½(𝑇uοΏ½(uοΏ½)π‘₯†)

≀ 𝐴2𝐡2uοΏ½(𝑇π‘₯†)

= 0

1Both assumptions are fulfilled in the quadratic Tikhonov case if the constraint set𝐢is weakly closed.

3.2 Discretization of Linear Inverse Problems It follows that𝑇 Μƒπ‘₯ = 𝑦†. Uniqueness ofπ‘₯†andβ„›( Μƒπ‘₯) ≀ℛ(π‘₯†)then implies that Μƒπ‘₯ = π‘₯†. Therefore, as claimed, Μ„π‘₯ ⇀ π‘₯†. In particular,βŸ¨πœ‰β€ , Μ„π‘₯ βˆ’ π‘₯β€ βŸ© β†’ 0. Due to (3.18),

uοΏ½β„›( Μ„π‘₯, π‘₯†) = β„›( Μ„π‘₯) βˆ’β„›(π‘₯†) βˆ’ βŸ¨πœ‰β€ , Μ„π‘₯ βˆ’ π‘₯β€ βŸ© β†’ 0.

This shows the first part.

For the second part, let𝛾 β‰₯ 0be arbitrary. The variational smoothness assumption, the inequality (3.17) and the fact that Μ„π‘₯ ∈ 𝐢yield

𝛽uοΏ½β„›( Μ„π‘₯, π‘₯†) ≀ β„›( Μ„π‘₯) βˆ’β„›(π‘₯†) + πœ“(uοΏ½(𝑇 Μ„π‘₯))

= 𝛾(β„›(π‘₯†) βˆ’β„›( Μ„π‘₯)) + (1 + 𝛾)(β„›( Μ„π‘₯) βˆ’β„›(π‘₯†)) + πœ“(uοΏ½(𝑇 Μ„π‘₯))

≀ 𝛾(β„›(π‘₯†) βˆ’β„›( Μ„π‘₯)) + 1 + 𝛾𝛼 (uοΏ½(π‘‡β„Žπ‘₯†) βˆ’uοΏ½(π‘‡β„Ž Μ„π‘₯)) + πœ“(uοΏ½(𝑇 Μ„π‘₯)).

Using (3.12), we obtain

uοΏ½(π‘‡β„Žπ‘₯†) βˆ’uοΏ½(π‘‡β„Ž Μ„π‘₯) =uοΏ½(π‘‡β„Žπ‘₯†) βˆ’uοΏ½(𝑦†) +uοΏ½(𝑦†) βˆ’uοΏ½(π‘‡β„Ž Μ„π‘₯)

≀ 𝐴(uοΏ½(π‘‡β„Žπ‘₯†) + πœ€uοΏ½) βˆ’ 1𝐴uοΏ½(π‘‡β„Ž Μ„π‘₯) + πœ€uοΏ½

≀ 𝐴𝐡ℛ(π‘₯†)πœ‚β„Ž+ (1 + 𝐴)πœ€uοΏ½βˆ’ 1𝐴𝐡uοΏ½(𝑇 Μ„π‘₯) +πœ‚β„Ž 𝐴 β„›( Μ„π‘₯).

We now choose𝛾 such that

πœ‚β„Ž

𝐴𝛼(1 + 𝛾) = 𝛾

to makeβ„›( Μ„π‘₯)-terms vanish. Note that we must have𝐴𝛼 > πœ‚β„Žfor𝛾 to be non-negative, which will however be ensured by the choice for𝛼below. We arrive at

𝛽uοΏ½β„›( Μ„π‘₯, π‘₯†) ≀ 1 + 𝛾𝛼 𝒆𝒓𝒓 βˆ’1 + 𝛾

𝐴𝐡𝛼 uοΏ½(𝑇 Μ„π‘₯) + πœ“(uοΏ½(𝑇 Μ„π‘₯)).

By definition of the conjugate function ofβˆ’πœ“, this can be estimated further by 𝛽uοΏ½β„›( Μ„π‘₯, π‘₯†) ≀ 1 + 𝛾𝛼 𝒆𝒓𝒓 +(βˆ’πœ“)βˆ—(βˆ’1 + 𝛾𝐴𝐡𝛼 ).

Setting𝑠 ∢= βˆ’(1 + 𝛾)(𝐴𝐡𝛼)βˆ’1, the infimum of the right hand side is

infuοΏ½<0(βˆ’π‘ π΄π΅ 𝒆𝒓𝒓 +(βˆ’πœ“)βˆ—(𝑠)) = βˆ’(βˆ’πœ“)βˆ—βˆ—(𝐴𝐡 𝒆𝒓𝒓) = πœ“(𝐴𝐡 𝒆𝒓𝒓) ≀ π΄π΅πœ“(𝒆𝒓𝒓).

Due to the second statement in Lemma 3.11, it is attained if 𝑠 ∈ πœ•(βˆ’πœ“)(𝐴𝐡 𝒆𝒓𝒓)

which is equivalent to

𝐴𝛼 ∈ πœ‚β„Ž+ (βˆ’π΅πœ•(βˆ’πœ“)(𝐴𝐡 𝒆𝒓𝒓))βˆ’1. (3.19) In particular, sinceπœ“is strictly increasing,𝐴𝛼 > πœ‚β„Žas needed above.

Remark3.13. 1. In the unconstrained case,infu�∈uοΏ½uοΏ½(π‘‡β„Žπ‘₯) = 12βˆ₯(1βˆ’π‘„β„Ž)𝑦uοΏ½βˆ₯2is attained ifπ‘‡β„Žπ‘₯ = π‘„β„Žπ‘¦uοΏ½. Therefore, by the characterization of the best approximation as orthogonal projection,

uοΏ½(π‘‡β„Žπ‘₯†) βˆ’ infu�∈uοΏ½uοΏ½(π‘‡β„Žπ‘₯) = 12βˆ₯π‘„β„Žπ‘‡π‘ƒβ„Žπ‘₯β€ βˆ’ π‘„β„Žπ‘¦uοΏ½βˆ₯2

≀ βˆ₯π‘„β„Ž(𝑦uοΏ½βˆ’ 𝑦†)βˆ₯2+ βˆ₯π‘„β„Žπ‘‡(1 βˆ’ π‘ƒβ„Ž)π‘₯†βˆ₯2

≀ βˆ₯𝑦uοΏ½βˆ’ 𝑦†βˆ₯2+ βˆ₯𝑇(1 βˆ’ π‘ƒβ„Ž)βˆ₯2βˆ₯(1 βˆ’ π‘ƒβ„Ž)π‘₯†βˆ₯2. Hence, (3.16) is implied by the conditions (3.10) of Theorem 3.7.

2. Since the explicit formπ‘‡β„Ž= π‘„β„Žπ‘‡π‘ƒβ„Žwas not used anywhere but just the fact that

βˆ₯𝑇 βˆ’ π‘‡β„Žβˆ₯ β†’ 0, the theorem actually gives an error estimate for more general operator errors. These may occur for example due to uncertainties in modeling the forward operator. Thus, the theorem can also be viewed as a way to partially alleviate some of the modeling issues discussed in Chapter 2.

3. Using the assumptions of Theorem 3.7 together with Theorem 3.9, we have shown a convergence rate

βˆ₯ Μ„π‘₯ βˆ’ π‘₯†βˆ₯ =uοΏ½((𝛿 + βˆ₯𝑇 βˆ’ π‘‡β„Žβˆ₯)2uοΏ½+12uοΏ½ ),

which is identical to the one in (3.11) with respect to the measurement error β€” both are of optimal order β€”, but somewhat worse with respect to the operator error.

4. Since the proof does not rely on linearity of𝑇andπ‘‡β„Ž, it actually also holds in the non-linear case. A very similar result β€” with almost identical proof β€” can be found in [LF12, Theorem 3.1], where convergence is shown under the assumption

πœ‚β„ŽβˆΌ sup

u�∈uοΏ½βˆ₯𝑇(π‘₯) βˆ’ π‘‡β„Ž(π‘₯)βˆ₯2 β†’ 0 (3.20)

for data fidelity terms given by a Banach space norm. Compared to that result, an error estimate of the form (3.12), despite having been introduced here merely as a notation, has been shown to be applicable in much more general settings than error assumptions based on norm-distances (cf. [HW13]). Moreover, (3.12) also holds in the linear case for unbounded constraint sets𝐢, when (3.20) is typically not satisfied.

4 A Penalty for ODFs

4.1 Choice of the Regularization Functional

In Chapter 3, Tikhonov regularization was introduced as a way to incorporate prior knowledge into the reconstruction, essentially by requiring that the norm of the re-construction be sufficiently small. It is clear that in this regard the space𝑋of possible solutions has to be chosen appropriately in order to achieve good reconstruction results.

For example, in order to enforce smoothness of the solutions, one might take𝑋 to be the Sobolev space𝐻1of weakly differentiable functions.

For ODF reconstruction, this choice is usually too restrictive, since it promotes isotropic smoothness of the reconstruction, i.e. in each spatial direction. ODFs on the other hand can only be assumed to be smooth along the fibers, but not perpendicular to them.

In [Dui05], Duits introduced a formalism for smoothing a related class of functions, namely functions onℝ2×𝑆1, with the aim to enhance elongated and crossing structures.

The idea is to identifyℝ2Γ— 𝑆1with the two-dimensional Euclidean groupSE(2)of translations and rotations. The group structure then provides

β€’ a distinct class of curves inSE(2), theexponential curvesparametrized by a starting point inSE(2)and a starting tangential vector, that can be viewed as generaliza-tions of straight lines toSE(2)and serve as a local model for elongated structures, and

β€’ a set of derivatives along these curves calledleft-invariant derivatives.

These derivatives are then used to construct non-linear diffusion filters onSE(2)by taking an image as initial stateπœ“|uοΏ½=0∢ SE(2) β†’ ℝof a diffusion equation

πœ•uοΏ½πœ“ = βˆ‡uοΏ½u�𝐷[πœ“]βˆ‡uοΏ½πœ“,

where βˆ‡uοΏ½ is the left-invariant gradient and the diffusion tensor 𝐷[πœ“]∢ SE(2) β†’ ℝ3Γ—3 is adapted to the local structure of the image. This equation is then solved up to some stopping time𝑑 = 𝜏. The resulting imageπœ“|uοΏ½=uοΏ½ can be seen as a denoised and enhanced version of the original image. The result depends crucially on the way 𝐷[πœ“]is constructed fromπœ“. In general, one tries to direct diffusion along elongated structures in order to enhance coherence, but not perpendicular to them in order to preserve edges. Doing this relies on estimating the orientation of the local structure,

which is done by finding the exponential curve that locally fits best to the image, in some appropriate sense.

In [DF11], Duits and Franken generalize their formalism to the three-dimensional case, i.e. to functions onℝ3×𝑆2, and apply this to the enhancement of HARDI reconstructions.

The main challenge here is the fact thatℝ3Γ— 𝑆2can not be identified withSE(3), the latter being isomorphic toℝ3⋉ SO(3). Instead,ℝ3Γ— 𝑆2can only be identified with a quotient ofSE(3)by a non-normal subgroup isomorphic toSO(2)and therefore does not inherit a group structure. To carry over concepts from theℝ2Γ— 𝑆1-case, one has to require that they are well-defined on the quotient.

The constructed diffusion filters are employed as a post-processing step, after recon-structing the ODF from DW-MRI data. Our aim is to include concepts from this formalism directly in the reconstruction algorithm by formulating a suitable penalty functional that can be used in for Tikhonov regularization.

In the following, a short overview of the𝑆𝐸(3)-formalism, in particular of the construc-tion of exponential curves, will be given.

Definition 4.1. The Euclidean Motion Group SE(𝑛)in 𝑛 dimensions is the semi-direct productℝuοΏ½ and the groupSO(𝑛)of rotations in𝑛dimensions,

SE(𝑛) ∢= ℝuοΏ½ β‹Š SO(𝑛),

i.e. SE(𝑛) = ℝuοΏ½ Γ— SO(𝑛)as a set, with the product between𝑔 = (𝑏, 𝑅) and𝑔′ = (𝑏′, 𝑅′) given by𝑔𝑔′ = (𝑏 + 𝑅𝑏′, 𝑅𝑅′). The unit element is𝑒 = (0, πŸ™)and the inverse of𝑔isπ‘”βˆ’1 = (βˆ’π‘…βˆ’1𝑏, π‘…βˆ’1).SE(𝑛)acts onℝuοΏ½ asgroup of rotations and translationsby𝑔π‘₯ ∢= 𝑅π‘₯ + 𝑏 forπ‘₯ ∈ ℝuοΏ½.

A matrix representation ofSE(𝑛)is given by

SE(𝑛) βˆ‹ (𝑏, 𝑅) ↦ (𝑅 𝑏0 1) ∈ ℝ(uοΏ½+1)Γ—(uοΏ½+1). SE(𝑛)acts in a natural way onℝuοΏ½Γ— 𝑆uοΏ½βˆ’1 by

(𝑏, 𝑅)(π‘₯, 𝑒) ∢= (𝑅π‘₯ + 𝑏, 𝑅𝑒), (4.1) i.e. by translating the spatial part and simultaneously rotating the spatial and the orientational parts. By identifying

ℝuοΏ½Γ— 𝑆uοΏ½βˆ’1 βˆ‹ (π‘₯, 𝑒) ↦ (𝑒 π‘₯0 1) ∈ ℝ(uοΏ½+1)Γ—2,

this action is given simply by matrix multiplication. Moreover, the tangent spaces to ℝuοΏ½Γ— 𝑆uοΏ½βˆ’1 can be identified with matrices of the form

𝑇(uοΏ½,uοΏ½)(ℝuοΏ½Γ— 𝑆uοΏ½βˆ’1) ≃ 𝑇u�ℝuοΏ½Γ— 𝑇u�𝑆uοΏ½βˆ’1 βˆ‹ (𝑋uοΏ½, 𝑋uοΏ½) ↦ (𝑋uοΏ½ 𝑋uοΏ½

0 0 ) ∈ ℝ(uοΏ½+1)Γ—2,

4.1 Choice of the Regularization Functional and with this identification, the derivative of the group action in (4.1) is again simply given by matrix multiplication.

We are particularly interested in the case𝑛 = 3. For this, it is convenient to introduce some additional notation:

Definition 4.2. Forπ‘Ž ∈ ℝ3, denote byπ‘ŽΓ— ∈ ℝ3Γ—3 the skew-symmetric matrix that acts as π‘ŽΓ—π‘₯ = π‘Ž Γ— π‘₯, i.e. the vector product betweenπ‘Žandπ‘₯, forπ‘₯ ∈ ℝ3.

For allπ‘Ž, 𝑏 ∈ ℝ3and𝑅 ∈ SO(3), one has

β€’ π‘ŽΓ—π‘Γ— = π‘π‘ŽuοΏ½ βˆ’ (𝑏uοΏ½π‘Ž)πŸ™,

β€’ (π‘ŽΓ—π‘)Γ— = π‘π‘ŽuοΏ½ βˆ’ π‘Žπ‘uοΏ½ = π‘ŽΓ—π‘Γ—βˆ’ π‘Γ—π‘ŽΓ—, and

β€’ (π‘…π‘Ž)Γ— = π‘…π‘ŽΓ—π‘…uοΏ½.

Having identifiedSE(𝑛)with a subgroup ofGl(𝑛 + 1, ℝ), one can prove the following theorem.

Theorem 4.3. SE(𝑛)is a Lie Group. For𝑛 = 3, its Lie Algebra𝔰𝔒(3)consists of all matrices of the form

𝔰𝔒(3) βˆ‹ (𝑏, π‘ŽΓ—) ≑ (π‘ŽΓ— 𝑏

0 0) ∈ ℝ4Γ—4,

forπ‘Ž, 𝑏 ∈ ℝ3, and the Lie bracket is given by the commutator[𝑋1, 𝑋2] = 𝑋1𝑋2βˆ’π‘‹2𝑋1. More explicitly,

[(𝑏1, π‘ŽΓ—1), (𝑏2, π‘ŽΓ—2)] = (π‘Ž1Γ— 𝑏2+ 𝑏1Γ— π‘Ž2, (π‘Ž1Γ— π‘Ž2)Γ—).

𝔰𝔒(𝑛)is isomorphic to the space of left-invariant vector fields onSE(3), i.e. vector fields1 𝑉 ∈ 𝛀(𝑇SE(3))fulfilling𝑉(π‘”β„Ž) = 𝑔𝑉(β„Ž)for all𝑔, β„Ž ∈ SE(3), the isomorphism being

𝔰𝔒(𝑛) βˆ‹ 𝑋 ↦ (𝑔 ↦ 𝑔𝑋) ∈ 𝛀(𝑇SE(3)).

Proof. See [Bak02].

As already mentioned above,SE(3)is larger than the domainℝ3Γ— 𝑆2on which ODFs are defined. The two spaces are related by a projectionπœ‹u�∢ SE(3) β†’ ℝ3Γ— 𝑆2which is constructed by choosing an arbitrary

πœ‰ ≑ (0, πœ‰0) ∈ ℝ3Γ— 𝑆2 and defining

πœ‹uοΏ½(π‘₯, 𝑅) ∢= (π‘₯, 𝑅)πœ‰ = (π‘₯, π‘…πœ‰0).

1𝛀(𝑇SE(3)) denotes the space of smooth sections of the tangent bundle of SE(3), i.e. functions 𝑉 ∢ SE(3) β†’ 𝑇SE(3)such that for all𝑔 ∈ SE(3),𝑉(𝑔)is in the tangential space𝑇uοΏ½SE(3)ofSE(3)at𝑔.

The kernel of this projection is thestabilizer

𝑆uοΏ½ = {β„Ž ∈ SE(3)∢ β„Žπœ‰ = πœ‰} = {(0, 𝑅) ∈ SE(3)∢ π‘…πœ‰0 = πœ‰0} ≃ SO(2).

πœ‹uοΏ½ is surjective, soℝ3Γ— 𝑆2can be identified with the coset spaceSE(3)/𝑆uοΏ½. Note that the stabilizer can also be written as

𝑆uοΏ½ = {(0, exp(π‘ πœ‰0Γ—))∢ 𝑠 ∈ ℝ} = {exp(𝑠(0, πœ‰0Γ—))∢ 𝑠 ∈ ℝ}.

Moreover,𝑋 ∈ 𝔰𝔒(3)fulfillsπ‘‹πœ‰ = 0if and only if𝑋 = 𝑑(0, πœ‰0Γ—)for some𝑑 ∈ ℝ.

In [DF11], a right inverse toπœ‹uοΏ½, i.e. and embedding ofℝ3Γ— 𝑆2intoSE(3), is constructed using Euler angles onSO(3) and spherical coordinates on 𝑆2. Unfortunately, this embedding is ill-defined at the coordinate poles. More generally, there is no such (continuous) embedding: if πœ„u�∢ ℝ3 Γ— 𝑆2 β†’ SE(3) fulfilled πœ‹uοΏ½ ∘ πœ„uοΏ½(𝑝) = 𝑝 for all 𝑝, restricting it to {0} Γ— 𝑆2 would yield a continuous map 𝜌∢ 𝑆2 β†’ SO(3) such that 𝜌(𝑒) πœ‰0 = 𝑒for all𝑒 ∈ 𝑆2. Now take a vectorπœ‚0 βŸ‚ πœ‰0. Then𝜌(𝑒) πœ‚0βŸ‚ 𝑒, so𝑒 ↦ 𝜌(𝑒) πœ‚0 is a non-vanishing, continuous tangential vector field on𝑆2, contradicting the hairy ball theorem2.

An important role in theSE(3)-formalism is played by exponential curves.

Definition 4.4. Let𝑔 ∈ SE(3)and𝑋 ∈ 𝔰𝔒(3). The curve𝛾 fulfilling 𝛾′(𝑑) = 𝛾(𝑑)𝑋, 𝛾(0) = 𝑔

is calledexponential curvethrough𝑔with parameter𝑋.

Using the matrix representation of𝔰𝔒(3), exponential curves are simply given using the matrix exponential as

𝛾(𝑑) = 𝑔 exp(𝑑𝑋). (4.2)

The simplest way to construct exponential curves on the quotient is as follows: given (π‘₯, 𝑒) ∈ ℝ3×𝑆2and𝑋 ∈ 𝔰𝔒(3), take𝑔 ∈ πœ‹uοΏ½βˆ’1(π‘₯, 𝑒), construct the exponential curve (4.2) and project it back toℝ3Γ— 𝑆2usingπœ‹uοΏ½. The question is: under what circumstances does the resulting curve not depend on the arbitrary choice of𝑔 andπœ‰? For this to hold, we must have

π‘”β„Ž exp(𝑑𝑋)πœ‰ = 𝑔 exp(𝑑𝑋)πœ‰ for all𝑑 ∈ ℝandβ„Ž ∈ 𝑆uοΏ½. Writing this as

exp(βˆ’π‘‘π‘‹) exp(𝑠(0, πœ‰0Γ—)) exp(𝑑𝑋)πœ‰ = πœ‰

for𝑠 ∈ ℝ, differentiating by𝑠and𝑑and setting𝑠 = 𝑑 = 0shows that [(0, πœ‰0Γ—), 𝑋]πœ‰ = 0.

2The hairy ball theorem states that every continuous, tangential vector field on𝑆2has a zero.

4.1 Choice of the Regularization Functional With𝑋 = (𝑏, π‘ŽΓ—), this can be expanded as

((πœ‰0Γ— π‘Ž) Γ— πœ‰0 πœ‰0Γ— 𝑏

0 0 ) = 0,

soπ‘Ž, 𝑏 ∈ span{πœ‰0}. Now

πœ‹uοΏ½(𝑔 exp((π‘‘πœ‰0, π‘ πœ‰0Γ—))) = πœ‹uοΏ½(𝑔 exp(𝑑(πœ‰0, 0))) = (π‘₯ + 𝑑𝑒, 𝑒) (4.3) is thehorizontal line, which is therefore the only well-defined exponential curve on the quotient.

On the other hand, the construction above may have been too restrictive, since𝑋 ∈ 𝔰𝔒(3) and𝑔 ∈ πœ‹uοΏ½βˆ’1(π‘₯, 𝑒)were chosen fixed and independent of each other. A possible way to improve the situation is to choose the parameter𝑋of the exponential curve depending on the choice of𝑔, i.e. to consider

ℝ βˆ‹ 𝑑 ↦ 𝑔 exp(𝑑𝑋(𝑔))πœ‰ (4.4)

for𝑋 ∢ πœ‹uοΏ½βˆ’1(π‘₯, 𝑒) β†’ 𝔰𝔒(3). For this to be well-defined, we need π‘”β„Ž exp(𝑑𝑋(π‘”β„Ž))πœ‰ = 𝑔 exp(𝑑𝑋(𝑔))πœ‰

⟺ β„Ž exp(𝑑𝑋(π‘”β„Ž))β„Žβˆ’1πœ‰ = exp(𝑑𝑋(𝑔))πœ‰ (4.5) for allβ„Ž ∈ 𝑆uοΏ½. Differentiating by𝑑and setting𝑑 = 0yields

Adβ„Žπ‘‹(π‘”β„Ž)πœ‰ = 𝑋(𝑔)πœ‰, (4.6)

whereAdβ„Žπ‘‹ ∢= β„Žπ‘‹β„Žβˆ’1is theadjoint representation. Explicitly,Adβ„Ž(𝑏, π‘ŽΓ—) = (𝑅𝑏, (π‘…π‘Ž)Γ—) forβ„Ž = (0, 𝑅). In particular,Adβ„Ž(0, πœ‰0Γ—) = (0, πœ‰0Γ—)for allβ„Ž ∈ 𝑆uοΏ½.

Define an inner product on𝔰𝔒(3)by

⟨(𝑏1, π‘ŽΓ—1), (𝑏2, π‘ŽΓ—2)⟩ ∢= π‘ŽuοΏ½1π‘Ž2+ 𝑏uοΏ½1𝑏2.

ThenAdβ„Žleavesspan{(0, πœ‰0Γ—)}and its orthogonal complement invariant. 𝑋(𝑔)can be uniquely decomposed as𝑋(𝑔) = π‘‹βŸ‚(𝑔) + 𝑐(𝑔)(0, πœ‰0Γ—)for someπ‘βˆΆ πœ‹uοΏ½βˆ’1(π‘₯, 𝑒) β†’ ℝand π‘‹βŸ‚(𝑔) βŸ‚ (0, πœ‰0Γ—). Then equation (4.6) implies that

Adβ„Žπ‘‹βŸ‚(π‘”β„Ž) = π‘‹βŸ‚(𝑔). (4.7)

Putting this back into (4.5), we obtain the requirement

exp(π‘‘π‘‹βŸ‚(𝑔) + 𝑑𝑐(π‘”β„Ž)(0, πœ‰0Γ—))πœ‰ = exp(π‘‘π‘‹βŸ‚(𝑔) + 𝑑𝑐(𝑔)(0, πœ‰0Γ—))πœ‰.

Differentiating twice with respect to𝑑, setting𝑑 = 0and using(0, πœ‰0Γ—)πœ‰ = 0leads to (𝑐(π‘”β„Ž) βˆ’ 𝑐(𝑔))(0, πœ‰0Γ—)π‘‹βŸ‚(𝑔)πœ‰ = 0.

Writeπ‘‹βŸ‚ asπ‘‹βŸ‚(𝑔) = (𝑏(𝑔), π‘Ž(𝑔)Γ—), withπ‘Ž(𝑔) βŸ‚ πœ‰0 by construction, and assume first can be summarized by requiring that, without loss of generality,𝑋has to fulfill

Adβ„Žπ‘‹(π‘”β„Ž) = 𝑋(𝑔) for allβ„Ž ∈ 𝑆uοΏ½ (4.8) in order for (4.4) to be well-defined.

Condition (4.8) has a nice interpretation. First, note thatπœ‹uοΏ½βˆ’1(π‘₯, 𝑒) βˆ‹ 𝑔 ↦ 𝑔𝑋(𝑔)is a vector field tangential toSE(3). Composing this with the differential of the projection πœ‹uοΏ½ yields a function𝑔 ↦ 𝑔𝑋(𝑔)πœ‰ ∈ 𝑇(uοΏ½,uοΏ½)(ℝ3Γ— 𝑆2). Since

π‘”β„Žπ‘‹(π‘”β„Ž)πœ‰ = 𝑔𝑋(𝑔)β„Žπœ‰ = 𝑔𝑋(𝑔)πœ‰,

this function is in fact constant. Therefore, every𝑋 ∢ πœ‹uοΏ½βˆ’1(π‘₯, 𝑒) β†’ 𝔰𝔒(3)fulfilling (4.8) determines a tangential vector πœ‹(uοΏ½,uοΏ½)βˆ— 𝑋 ∈ 𝑇(uοΏ½,uοΏ½)(ℝ3 Γ— 𝑆2). πœ‹(uοΏ½,uοΏ½)βˆ— is surjective and uοΏ½(πœ‹(uοΏ½,uοΏ½)βˆ— ) = span{𝑔 ↦ (0, πœ‰0Γ—)}. Thus, the space of functions fulfilling (4.8) can be interpreted as the tangent space𝑇(uοΏ½,uοΏ½)(ℝ3Γ— 𝑆2)plus one additional dimension.

An inverse toπœ‹(uοΏ½,uοΏ½)βˆ— can be given explicitly: ifπœ‹(uοΏ½,uοΏ½)βˆ— 𝑋 = (𝑉uοΏ½, 𝑉uοΏ½), then

otherwise. The spatial part is a helix with axisπ‘Š, while the orientational part performs the corresponding rotation that keeps its components constant with respect to the moving frame of the spatial part. In particular, the curves actually depend on𝑐 ∈ ℝ, so the space of exponential curves at (π‘₯, 𝑒) is larger than the tangential space 𝑇(uοΏ½,uοΏ½)(ℝ3Γ— 𝑆2).

4.1 Choice of the Regularization Functional In [Fra08], orientation estimation inSE(2)is done by introducingleft-invariant deriva-tives, which are essentially derivatives along exponential curves. Since these curves are parametrized by elements in𝔰𝔒(2), the left-invariant gradientβˆ‡uοΏ½πœ“of a function πœ“βˆΆ SE(2) β†’ ℝcan be interpreted as a map fromSE(2)to the fixed space𝔰𝔒(2)βˆ—. There-fore, it is possible to take second derivatives, which is in general not possible for the usual derivative. In particular, it is possible to compute theleft-invariant Hessian matrix.

The local orientation is then taken as the (right) singular vector corresponding to the smallest singular value.3 This is motivated by the requirement thatβˆ‡uοΏ½πœ“should change as little as possible along the exponential curve corresponding to the local orientation 𝑋(𝑔)at𝑔 ∈ SE(2):

𝑋(𝑔) = argmin

βˆ₯uοΏ½βˆ₯=1 βˆ₯πœ•uοΏ½βˆ‡uοΏ½πœ“(𝑔𝛾uοΏ½(𝑑))∣uοΏ½=0βˆ₯. (4.9) When applying this to our case, several problems arise:

β€’ There are too many exponential curves: they are parametrized by a six-dimen-sional space, while ODFs are only defined onℝ3Γ— 𝑆2. Intuitively, it seems clear that local knowledge of an ODF can not determine a unique exponential curve.

Indeed, the left-invariant Hessian in this case is singular withleftsingular vector πœ‰0Γ—. The interpretation of the corresponding rightsingular vector is not clear.

Moreover, numerical experiments showed frequent cases in which there is more than one small singular value; in these cases it is not clear which singular value is the β€œtrivial” one and should be discarded.

β€’ On the other hand, restricting the space of allowed curves in the most obvious way β€” namely putting 𝑐 = 0 in the definition of π‘Š above β€” does not seem reasonable: it restricts the spatial part𝛾uοΏ½ of the curve to be a helix with axis perpendicular to the orientational part𝑒of the starting point(π‘₯, 𝑒) ∈ ℝ3Γ— 𝑆2. In addition to being a rather arbitrary choice, it also has the undesirable effect that it is impossible for all points in{π‘₯} Γ— π‘ˆ βŠ‚ ℝ3Γ— 𝑆2withπ‘ˆ βŠ‚ 𝑆2an arbitrarily small, open set to belong to the same oriented structure except for a straight line, π‘Š = 0.

β€’ A more sensible approach might be to restrict some intrinsic parameters of the projected exponential curve. For example, one might assume that the spatial part of the curve is torsion-free. The torsion of𝛾uοΏ½ can be explicitly calculated to be𝜏 = π‘Š β‹… 𝑉uοΏ½, leading to an additional indefinite quadratic constraint in the SE(3)-version of (4.9), which is therefore more complicated and time-intensive to solve, and may not even have a unique solution.

β€’ In order to estimate local orientation, a (small) minimization problem like (4.9) has to be solved at each point of a discretization ofℝ3 Γ— 𝑆2. While this may be acceptable when solving a diffusion equation as a post-processing step, for example by computing one low-dimensional SVD at each point for each time

3Using singular vectors instead of eigenvectors is necessary since the Hessian is in general

3Using singular vectors instead of eigenvectors is necessary since the Hessian is in general