3.1.2 Constrained Tikhonov Regularization
In some cases, one knows in advance that that the true solutionπ₯β lies in some convex subset πΆ β π. In those cases, the result can often be improved by replacing the Tikhonov functionalπ½uοΏ½in (3.7) with
π½uοΏ½uοΏ½(π₯) βΆ= β₯ππ₯ β π¦β₯2+ πΌβ₯π₯β₯2+ πuοΏ½(π₯), where
πuοΏ½(π₯) =β§{
β¨{β©0, π₯ β πΆ,
β, otherwise is the indicator function of the setπΆ. Thus, one solves
π uοΏ½,uοΏ½π¦uοΏ½ = argmin
uοΏ½βuοΏ½ π½uοΏ½(π₯).
IfπΆis closed and convex, a minimizer ofπ½uοΏ½uοΏ½exists and is uniquely determined (see for example [EHN96, Theorem 5.15]). Sinceπ uοΏ½,uοΏ½is non-linear, spectral theory can not be employed in the convergence analysis. Convergence of orderβπΏfor this method was shown in [Neu88] under theprojected source condition
π₯β = πuοΏ½(πβπ) for someπ β π, (3.9) whereπuοΏ½βΆ π β πis the metric projection ontoπΆ. This condition is similar to a HΓΆlder source conditionπ₯β ββ((πβπ)1/2) =β(πβ)with exponent Β½, but can be signiο¬cantly weaker depending on the constraint setπΆ.
3.2 Discretization of Linear Inverse Problems
In practice, the problem (3.1) can only be solved in ο¬nite dimensions. Therefore, the continuous formulation has to be discretized. This is done by introducing orthogonal projectionsπβ ββ(π)andπβ ββ(π)with ο¬nite dimensional range. Here,β > 0is a discretization parameter. Then (3.1) is replaced by
πβπ₯ βΆ= πβππβπ₯ = πβπ¦.
This equation is then solved forπ₯ β β(πβ)by the methods introduced in the previ-ous sections. The aim of this section is to investigate the convergence of discretized Tikhonov regularization under a convex constraint.
In the following, it is assumed thatβ₯πββ πβ₯ β 0asβ β 0. To check this property in concrete cases, the following lemma may be useful.
Lemma 3.6. The following statements are equivalent:
1. β₯πββ πβ₯ β 0asβ β 0
2. β₯π(1 β πβ)β₯ β 0andβ₯(1 β πβ)πβ₯ β 0
3. πis compact,πβ β 1pointwise onβ(πβ)andπββ 1pointwise onβ(π) Proof. (1.) implies (2.) since
β₯(1 β πβ)πβ₯ β€ β₯πββ πβπβ₯ + β₯πββ πβ₯ β€ 2β₯πββ πβ₯, and similarly forβ₯π(1 β πβ)β₯. The reverse implication holds due to
β₯πββ πβ₯ β€ β₯(1 β πβ)ππββ₯ + β₯π(1 β πβ)β₯.
For the rest, note that due toβ₯π(1βπβ)β₯ = β₯(1βπβ)πββ₯, it suο¬ces to show the statements forπβ. The implication (2.) β (3.) holds sinceπ is the limit of ο¬nite-dimensional operatorsπβπ and hence compact, and sinceπβπ¦ β π¦for allπ¦ = ππ₯withπ₯ β π. To see the reverse, assume thatβ₯(1 β πβ)πβ₯ β 0. Then there areπ > 0and sequences (βuοΏ½)uοΏ½ββ β βand(π₯uοΏ½)uοΏ½ββ β πwithβuοΏ½ β 0,β₯π₯uοΏ½β₯ = 1and
β₯(1 β πβuοΏ½)ππ₯uοΏ½β₯ β₯ π.
By compactness, we can assume without loss of generality that ππ₯uοΏ½ β π¦ β β(π). Then the previous inequality implies that(1 β πβ)π¦ β 0. On the other hand, uniform boundedness of (πβ)β>0 implies that πβ β 1 pointwise also on β(π), which is a contradiction.
We cite the following convergence result for the discretized solution under HΓΆlder source conditions in the unconstrained case from [PV90].
Theorem 3.7. Let πβ,πβ and πβ be as above, π¦ β β(π) and π¦uοΏ½ β π with β₯π¦ β π¦uοΏ½β₯ β€ πΏ. Assume thatβ₯π β πββ₯ β 0asβ β 0, and that the regularization parameterπΌ > 0is chosen such that
πΌ β 0, πΏπΌ β 0,2 β₯π(1 β πβ)β₯2
πΌ β€ π (3.10)
asπΏ, β β 0for someπ > 0. Denote by Μπ₯ ββ(πβ)the discrete Tikhonov-regularized solution fulο¬lling
(πββπβ+ πΌ) Μπ₯ = πββπ¦uοΏ½.
Then uοΏ½,ββ0lim Μπ₯ = π₯β .
If, in addition,
π₯β ββ((πβπ)uοΏ½)
3.2 Discretization of Linear Inverse Problems for some0 < π β€ 12, and
π1πΌ β€ πΏ2uοΏ½+12 + β₯π(1 β πβ)β₯2 β€ π2πΌ for someπ1, π2> 0, then
β₯ Μπ₯ β π₯β β₯ =uοΏ½(πΏ2uοΏ½+12uοΏ½ + β₯π β πββ₯2uοΏ½) (3.11) asπΏ, β β 0.
Unfortunately, the proof in [PV90] relies on spectral theory and therefore does not generalize to the constrained setting.
To address this problem, spectral theory can be replaced by variational techniques in regularization theory developed in the last decade. They were ο¬rst introduced in [HKPS07] for problems with non-linear, non-smooth operators, but also turned out useful to handle more general data ο¬delities and penalties as well as constrained problems. In the following, we will prove a convergence result similar to Theorem 3.7 using these techniques.
Even though we are only interested in the Tikhonov case, the convergence theorem will be presented in a more abstract way, for which some additional notation is needed.
Assume that we are given an ideal data ο¬delity functional uοΏ½ β‘ uοΏ½uοΏ½β βΆ π β [0, β]
with
uοΏ½(π¦) = 0 βΊ π¦ = π¦β ,
and anempirical data ο¬delity functionaluοΏ½ β‘uοΏ½uοΏ½uοΏ½βΆ π β β βͺ {β}, both convex, which are connected by the error assumption
π΄1 uοΏ½(π¦) β πuοΏ½ β€uοΏ½(π¦) βuοΏ½(π¦β ) β€ π΄(uοΏ½(π¦) + πuοΏ½) βπ¦ β π
for some π΄ β₯ 1and πuοΏ½ > 0 such that πuοΏ½ β 0 asπΏ β 0. The convexregularization penaltyis denoted byββΆ π β β βͺ {β}, and the discretization error is controlled by the assumption
π΅1 uοΏ½(ππ₯) β πββ(π₯) β€uοΏ½(πβπ₯) β€ π΅(uοΏ½(ππ₯) + πββ(π₯)) βπ₯ β πΆ β π (3.12a) forπ΅ β₯ 1andπβ> 0withπββ 0asβ β 0, and whereπΆ β πis the convex constraint set. Then (generalized) discrete constrained Tikhonov regularization consists in solving the optimization problem
Μπ₯ β argmin
uοΏ½βuοΏ½ (uοΏ½(πβπ₯) + πΌβ(π₯)). (3.13)
In the variational formalism, convergence is usually investigated in terms of the Breg-man distanceof the convex regularization penalty. It is deο¬ned as
uοΏ½β(π₯, π₯β ) β‘uοΏ½uοΏ½ββ (π₯, π₯β ) =β(π₯) ββ(π₯β ) β β¨πβ , π₯ β π₯β β©,
whereπβ β πβ(π₯β )is a ο¬xed element in the subgradient of β at π₯β . Note that in general,uοΏ½β is neither symmetric nor does it satisfy a triangle inequality.
An important part in the variational approach consists in replacing the spectral source condition (3.5) by a condition in the form of a variational inequality.
Deο¬nition 3.8. π₯β β πis said to satisfy an (additive)variational smoothness assumption if there areπ½ > 0and a concave index functionπβΆ β+ β βwith
π½uοΏ½β(π₯, π₯β ) β€β(π₯) ββ(π₯β ) + π(uοΏ½(ππ₯)) βπ₯ β πΆ. (3.14) Quadratic Tikhonov regularization is the special case
uοΏ½(π¦) = 12β₯π¦ β π¦β β₯2, uοΏ½(π¦) = 12β₯π¦ β π¦uοΏ½β₯2, and
β(π₯) = 12β₯π₯β₯2+ πuοΏ½(π₯).
(3.15)
In this case, assuming thatπ₯β β πΆ, the Bregman distance atπ₯ β πΆis simply uοΏ½β(π₯, π₯β ) = 12β₯π₯ β π₯β β₯2+ πuοΏ½(π₯).
To show that (3.12) is fulο¬lled case withπ΄ = 2,π΅ = 1and πuοΏ½ = β₯π¦uοΏ½β π¦β β₯2 and πβ= β₯π β πββ₯2, we use the inequalityβ₯π + πβ₯2β€ 2β₯πβ₯2+ 2β₯πβ₯2to obtain
12uοΏ½(π¦) β πuοΏ½ β€ 12β₯π¦ β π¦uοΏ½β₯2β 12β₯π¦β β π¦uοΏ½β₯2=uοΏ½(π¦) βuοΏ½(π¦β ) and
uοΏ½(π¦) βuοΏ½(π¦β ) β€ β₯π¦ β π¦β β₯2+ 12β₯π¦uοΏ½β π¦β β₯2β€ 2(uοΏ½(π¦) + πuοΏ½).
The inequality (3.12a) follows from
uοΏ½(πβπ₯) = 12β₯πβπ₯ β π¦β β₯2
β€ β₯ππ₯ β π¦β β₯2+ β₯(π β πβ)π₯β₯2
β€ β₯ππ₯ β π¦β β₯2+ β₯π β πββ₯2β₯π₯β₯2
=uοΏ½(ππ₯) + πββ(π₯).
3.2 Discretization of Linear Inverse Problems The inequalityuοΏ½(ππ₯) β€uοΏ½(πβπ₯) + πββ(π₯)is obtained in the same way by exchanging πandπβ.
Several results on the relations between spectral source conditions and variational smoothness assumptions have been obtained (an overview is given in [Fle13]). The following is from [HY10].
Theorem 3.9. In the quadratic Tikhonov case(3.15)and for a linear operatorπ βΆ π β π, the HΓΆlder source condition(3.5)withπ = πuοΏ½ given by(3.6)for0 < π β€ 12 implies(3.14)with π βΌ πuοΏ½forπ = 2uοΏ½+12uοΏ½ andπ½ β (0, 1)arbitrary.
For larger HΓΆlder exponents, this theorem can not hold. In fact, in [HY10] it is also proved that (3.14) withπ βΌ πuοΏ½ forπ > 1/2impliesπ₯β = 0. So in this sense, variational smoothness assumptions only cover HΓΆlder source conditions withπ β€ 1/2.
Moreover, in [FH11], equivalence between the projected source condition (3.9) and (3.14) withπ(π‘) βΌ βπ‘was shown. This serves as the main motivation to approach the discrete, constrained problem with these methods.
To prove the convergence theorem, some basic results from convex analysis are re-quired:
Deο¬nition 3.10. Given a convex, lower semi-continuous functionπ βΆ βuοΏ½ β β βͺ {β}not constantlyβ, itsconjugate functionπβ is deο¬ned by
πβ(π¦) = sup
uοΏ½ββuοΏ½(π¦uοΏ½π₯ β π (π₯)), π β β.
Ifπ is only deο¬ned on a subset ofβuοΏ½, it is understood to be extended to all of it byβ. Lemma 3.11. The following statements hold:
β’ π (π₯) + πβ(π¦) β₯ π¦uοΏ½π₯,
β’ π (π₯) + πβ(π¦) = π¦uοΏ½π₯ βΊ π¦ β ππ (π₯) βΊ π₯ β ππβ(π¦), and
β’ πββ βΆ= (πβ)β = π.
Proof. More general versions of these statements are shown for example in [ET76, Chapter 1].
Theorem 3.12. Let the relations(3.12) hold and Μπ₯be deο¬ned as the assumed to be unique minimizer in(3.13). Assume further thatπ₯β is the uniqueβ-minimizing solution toππ₯ = π¦β .
1. LetuοΏ½ and β be weakly lower semi-continuous and the sets {π₯ β π βΆ β(π₯) β€ π} be weakly sequentially compact for allπ > 0.1 IfπΌis chosen such thatπΌ β 0and
πΌ(1 uοΏ½(πβπ₯β ) β infuοΏ½βuοΏ½uοΏ½(πβπ₯)) β 0 (3.16) asπΏ, β β 0, then
uοΏ½,ββ0limuοΏ½β( Μπ₯, π₯β ) = 0.
2. If the variational smoothness assumption(3.14)holds true, then π½uοΏ½β( Μπ₯, π₯β ) β€ π΄π΅π(πππ)
with πππ = (π΄π΅ + π΄β1)β(π₯β )πβ+ (1 + π΄)πuοΏ½, ifπΌis chosen suitably (see(3.19)below).
Proof. By deο¬nition,
uοΏ½(πβ Μπ₯) + πΌβ( Μπ₯) β€uοΏ½(πβπ₯β ) + πΌβ(π₯β ) (3.17) and thus, using (3.16),
lim sup
uοΏ½,ββ0 β( Μπ₯) β€β(π₯β ). (3.18)
We now show that each sequence of minimizers( Μπ₯uοΏ½)uοΏ½ββ fulο¬lling the requirements with corresponding errorsπuοΏ½, πuοΏ½ β 0has a subsequence that converges weakly toπ₯β . This then implies that Μπ₯ β π₯β asπΏ, β β 0.
Due to (3.18) and the compactness assumption onβ,( Μπ₯uοΏ½)has a weakly convergent subsequence uοΏ½(uοΏ½)Μπ₯ β Μπ₯ asπ β β. Thenπ Μπ₯uοΏ½(uοΏ½) β Μππ₯, since every continuous linear operator is weakly continuous. Using the weak lower semi-continuity ofuοΏ½, (3.17) and the error relations (3.12), we obtain
uοΏ½(π Μπ₯) β€ lim infuοΏ½ββ uοΏ½(π Μπ₯uοΏ½(uοΏ½))
β€ lim infuοΏ½ββ π΅(uοΏ½(πuοΏ½(uοΏ½) uοΏ½(uοΏ½)Μπ₯ ) + πββ(π₯))
β€ lim infuοΏ½ββ π΅uοΏ½(πuοΏ½(uοΏ½) uοΏ½(uοΏ½)Μπ₯ )
β€ lim infuοΏ½ββ π΄π΅(uοΏ½uοΏ½(uοΏ½)(πuοΏ½(uοΏ½) uοΏ½(uοΏ½)Μπ₯ ) βuοΏ½uοΏ½(uοΏ½)(π¦β ))
β€ lim infuοΏ½ββ π΄π΅(uοΏ½uοΏ½(uοΏ½)(πuοΏ½(uοΏ½)π₯β ) βuοΏ½uοΏ½(uοΏ½)(π¦β ) + πΌuοΏ½(uοΏ½)(β(π₯β ) ββ( Μπ₯uοΏ½(uοΏ½))))
= lim infuοΏ½ββ π΄π΅(uοΏ½uοΏ½(uοΏ½)(πuοΏ½(uοΏ½)π₯β ) βuοΏ½uοΏ½(uοΏ½)(π¦β ))
β€ lim infuοΏ½ββ π΄2π΅uοΏ½(πuοΏ½(uοΏ½)π₯β )
β€ π΄2π΅2uοΏ½(ππ₯β )
= 0
1Both assumptions are fulο¬lled in the quadratic Tikhonov case if the constraint setπΆis weakly closed.
3.2 Discretization of Linear Inverse Problems It follows thatπ Μπ₯ = π¦β . Uniqueness ofπ₯β andβ( Μπ₯) β€β(π₯β )then implies that Μπ₯ = π₯β . Therefore, as claimed, Μπ₯ β π₯β . In particular,β¨πβ , Μπ₯ β π₯β β© β 0. Due to (3.18),
uοΏ½β( Μπ₯, π₯β ) = β( Μπ₯) ββ(π₯β ) β β¨πβ , Μπ₯ β π₯β β© β 0.
This shows the ο¬rst part.
For the second part, letπΎ β₯ 0be arbitrary. The variational smoothness assumption, the inequality (3.17) and the fact that Μπ₯ β πΆyield
π½uοΏ½β( Μπ₯, π₯β ) β€ β( Μπ₯) ββ(π₯β ) + π(uοΏ½(π Μπ₯))
= πΎ(β(π₯β ) ββ( Μπ₯)) + (1 + πΎ)(β( Μπ₯) ββ(π₯β )) + π(uοΏ½(π Μπ₯))
β€ πΎ(β(π₯β ) ββ( Μπ₯)) + 1 + πΎπΌ (uοΏ½(πβπ₯β ) βuοΏ½(πβ Μπ₯)) + π(uοΏ½(π Μπ₯)).
Using (3.12), we obtain
uοΏ½(πβπ₯β ) βuοΏ½(πβ Μπ₯) =uοΏ½(πβπ₯β ) βuοΏ½(π¦β ) +uοΏ½(π¦β ) βuοΏ½(πβ Μπ₯)
β€ π΄(uοΏ½(πβπ₯β ) + πuοΏ½) β 1π΄uοΏ½(πβ Μπ₯) + πuοΏ½
β€ π΄π΅β(π₯β )πβ+ (1 + π΄)πuοΏ½β 1π΄π΅uοΏ½(π Μπ₯) +πβ π΄ β( Μπ₯).
We now chooseπΎ such that
πβ
π΄πΌ(1 + πΎ) = πΎ
to makeβ( Μπ₯)-terms vanish. Note that we must haveπ΄πΌ > πβforπΎ to be non-negative, which will however be ensured by the choice forπΌbelow. We arrive at
π½uοΏ½β( Μπ₯, π₯β ) β€ 1 + πΎπΌ πππ β1 + πΎ
π΄π΅πΌ uοΏ½(π Μπ₯) + π(uοΏ½(π Μπ₯)).
By deο¬nition of the conjugate function ofβπ, this can be estimated further by π½uοΏ½β( Μπ₯, π₯β ) β€ 1 + πΎπΌ πππ +(βπ)β(β1 + πΎπ΄π΅πΌ ).
Settingπ βΆ= β(1 + πΎ)(π΄π΅πΌ)β1, the inο¬mum of the right hand side is
infuοΏ½<0(βπ π΄π΅ πππ +(βπ)β(π )) = β(βπ)ββ(π΄π΅ πππ) = π(π΄π΅ πππ) β€ π΄π΅π(πππ).
Due to the second statement in Lemma 3.11, it is attained if π β π(βπ)(π΄π΅ πππ)
which is equivalent to
π΄πΌ β πβ+ (βπ΅π(βπ)(π΄π΅ πππ))β1. (3.19) In particular, sinceπis strictly increasing,π΄πΌ > πβas needed above.
Remark3.13. 1. In the unconstrained case,infuοΏ½βuοΏ½uοΏ½(πβπ₯) = 12β₯(1βπβ)π¦uοΏ½β₯2is attained ifπβπ₯ = πβπ¦uοΏ½. Therefore, by the characterization of the best approximation as orthogonal projection,
uοΏ½(πβπ₯β ) β infuοΏ½βuοΏ½uοΏ½(πβπ₯) = 12β₯πβππβπ₯β β πβπ¦uοΏ½β₯2
β€ β₯πβ(π¦uοΏ½β π¦β )β₯2+ β₯πβπ(1 β πβ)π₯β β₯2
β€ β₯π¦uοΏ½β π¦β β₯2+ β₯π(1 β πβ)β₯2β₯(1 β πβ)π₯β β₯2. Hence, (3.16) is implied by the conditions (3.10) of Theorem 3.7.
2. Since the explicit formπβ= πβππβwas not used anywhere but just the fact that
β₯π β πββ₯ β 0, the theorem actually gives an error estimate for more general operator errors. These may occur for example due to uncertainties in modeling the forward operator. Thus, the theorem can also be viewed as a way to partially alleviate some of the modeling issues discussed in Chapter 2.
3. Using the assumptions of Theorem 3.7 together with Theorem 3.9, we have shown a convergence rate
β₯ Μπ₯ β π₯β β₯ =uοΏ½((πΏ + β₯π β πββ₯)2uοΏ½+12uοΏ½ ),
which is identical to the one in (3.11) with respect to the measurement error β both are of optimal order β, but somewhat worse with respect to the operator error.
4. Since the proof does not rely on linearity ofπandπβ, it actually also holds in the non-linear case. A very similar result β with almost identical proof β can be found in [LF12, Theorem 3.1], where convergence is shown under the assumption
πββΌ sup
uοΏ½βuοΏ½β₯π(π₯) β πβ(π₯)β₯2 β 0 (3.20)
for data ο¬delity terms given by a Banach space norm. Compared to that result, an error estimate of the form (3.12), despite having been introduced here merely as a notation, has been shown to be applicable in much more general settings than error assumptions based on norm-distances (cf. [HW13]). Moreover, (3.12) also holds in the linear case for unbounded constraint setsπΆ, when (3.20) is typically not satisο¬ed.
4 A Penalty for ODFs
4.1 Choice of the Regularization Functional
In Chapter 3, Tikhonov regularization was introduced as a way to incorporate prior knowledge into the reconstruction, essentially by requiring that the norm of the re-construction be suο¬ciently small. It is clear that in this regard the spaceπof possible solutions has to be chosen appropriately in order to achieve good reconstruction results.
For example, in order to enforce smoothness of the solutions, one might takeπ to be the Sobolev spaceπ»1of weakly diο¬erentiable functions.
For ODF reconstruction, this choice is usually too restrictive, since it promotes isotropic smoothness of the reconstruction, i.e. in each spatial direction. ODFs on the other hand can only be assumed to be smooth along the ο¬bers, but not perpendicular to them.
In [Dui05], Duits introduced a formalism for smoothing a related class of functions, namely functions onβ2Γπ1, with the aim to enhance elongated and crossing structures.
The idea is to identifyβ2Γ π1with the two-dimensional Euclidean groupSE(2)of translations and rotations. The group structure then provides
β’ a distinct class of curves inSE(2), theexponential curvesparametrized by a starting point inSE(2)and a starting tangential vector, that can be viewed as generaliza-tions of straight lines toSE(2)and serve as a local model for elongated structures, and
β’ a set of derivatives along these curves calledleft-invariant derivatives.
These derivatives are then used to construct non-linear diο¬usion ο¬lters onSE(2)by taking an image as initial stateπ|uοΏ½=0βΆ SE(2) β βof a diο¬usion equation
πuοΏ½π = βuοΏ½uοΏ½π·[π]βuοΏ½π,
where βuοΏ½ is the left-invariant gradient and the diο¬usion tensor π·[π]βΆ SE(2) β β3Γ3 is adapted to the local structure of the image. This equation is then solved up to some stopping timeπ‘ = π. The resulting imageπ|uοΏ½=uοΏ½ can be seen as a denoised and enhanced version of the original image. The result depends crucially on the way π·[π]is constructed fromπ. In general, one tries to direct diο¬usion along elongated structures in order to enhance coherence, but not perpendicular to them in order to preserve edges. Doing this relies on estimating the orientation of the local structure,
which is done by ο¬nding the exponential curve that locally ο¬ts best to the image, in some appropriate sense.
In [DF11], Duits and Franken generalize their formalism to the three-dimensional case, i.e. to functions onβ3Γπ2, and apply this to the enhancement of HARDI reconstructions.
The main challenge here is the fact thatβ3Γ π2can not be identiο¬ed withSE(3), the latter being isomorphic toβ3β SO(3). Instead,β3Γ π2can only be identiο¬ed with a quotient ofSE(3)by a non-normal subgroup isomorphic toSO(2)and therefore does not inherit a group structure. To carry over concepts from theβ2Γ π1-case, one has to require that they are well-deο¬ned on the quotient.
The constructed diο¬usion ο¬lters are employed as a post-processing step, after recon-structing the ODF from DW-MRI data. Our aim is to include concepts from this formalism directly in the reconstruction algorithm by formulating a suitable penalty functional that can be used in for Tikhonov regularization.
In the following, a short overview of theππΈ(3)-formalism, in particular of the construc-tion of exponential curves, will be given.
Deο¬nition 4.1. The Euclidean Motion Group SE(π)in π dimensions is the semi-direct productβuοΏ½ and the groupSO(π)of rotations inπdimensions,
SE(π) βΆ= βuοΏ½ β SO(π),
i.e. SE(π) = βuοΏ½ Γ SO(π)as a set, with the product betweenπ = (π, π ) andπβ² = (πβ², π β²) given byππβ² = (π + π πβ², π π β²). The unit element isπ = (0, π)and the inverse ofπisπβ1 = (βπ β1π, π β1).SE(π)acts onβuοΏ½ asgroup of rotations and translationsbyππ₯ βΆ= π π₯ + π forπ₯ β βuοΏ½.
A matrix representation ofSE(π)is given by
SE(π) β (π, π ) β¦ (π π0 1) β β(uοΏ½+1)Γ(uοΏ½+1). SE(π)acts in a natural way onβuοΏ½Γ πuοΏ½β1 by
(π, π )(π₯, π’) βΆ= (π π₯ + π, π π’), (4.1) i.e. by translating the spatial part and simultaneously rotating the spatial and the orientational parts. By identifying
βuοΏ½Γ πuοΏ½β1 β (π₯, π’) β¦ (π’ π₯0 1) β β(uοΏ½+1)Γ2,
this action is given simply by matrix multiplication. Moreover, the tangent spaces to βuοΏ½Γ πuοΏ½β1 can be identiο¬ed with matrices of the form
π(uοΏ½,uοΏ½)(βuοΏ½Γ πuοΏ½β1) β πuοΏ½βuοΏ½Γ πuοΏ½πuοΏ½β1 β (πuοΏ½, πuοΏ½) β¦ (πuοΏ½ πuοΏ½
0 0 ) β β(uοΏ½+1)Γ2,
4.1 Choice of the Regularization Functional and with this identiο¬cation, the derivative of the group action in (4.1) is again simply given by matrix multiplication.
We are particularly interested in the caseπ = 3. For this, it is convenient to introduce some additional notation:
Deο¬nition 4.2. Forπ β β3, denote byπΓ β β3Γ3 the skew-symmetric matrix that acts as πΓπ₯ = π Γ π₯, i.e. the vector product betweenπandπ₯, forπ₯ β β3.
For allπ, π β β3andπ β SO(3), one has
β’ πΓπΓ = ππuοΏ½ β (πuοΏ½π)π,
β’ (πΓπ)Γ = ππuοΏ½ β ππuοΏ½ = πΓπΓβ πΓπΓ, and
β’ (π π)Γ = π πΓπ uοΏ½.
Having identiο¬edSE(π)with a subgroup ofGl(π + 1, β), one can prove the following theorem.
Theorem 4.3. SE(π)is a Lie Group. Forπ = 3, its Lie Algebraπ°π’(3)consists of all matrices of the form
π°π’(3) β (π, πΓ) β‘ (πΓ π
0 0) β β4Γ4,
forπ, π β β3, and the Lie bracket is given by the commutator[π1, π2] = π1π2βπ2π1. More explicitly,
[(π1, πΓ1), (π2, πΓ2)] = (π1Γ π2+ π1Γ π2, (π1Γ π2)Γ).
π°π’(π)is isomorphic to the space of left-invariant vector ο¬elds onSE(3), i.e. vector ο¬elds1 π β π€(πSE(3))fulο¬llingπ(πβ) = ππ(β)for allπ, β β SE(3), the isomorphism being
π°π’(π) β π β¦ (π β¦ ππ) β π€(πSE(3)).
Proof. See [Bak02].
As already mentioned above,SE(3)is larger than the domainβ3Γ π2on which ODFs are deο¬ned. The two spaces are related by a projectionπuοΏ½βΆ SE(3) β β3Γ π2which is constructed by choosing an arbitrary
π β‘ (0, π0) β β3Γ π2 and deο¬ning
πuοΏ½(π₯, π ) βΆ= (π₯, π )π = (π₯, π π0).
1π€(πSE(3)) denotes the space of smooth sections of the tangent bundle of SE(3), i.e. functions π βΆ SE(3) β πSE(3)such that for allπ β SE(3),π(π)is in the tangential spaceπuοΏ½SE(3)ofSE(3)atπ.
The kernel of this projection is thestabilizer
πuοΏ½ = {β β SE(3)βΆ βπ = π} = {(0, π ) β SE(3)βΆ π π0 = π0} β SO(2).
πuοΏ½ is surjective, soβ3Γ π2can be identiο¬ed with the coset spaceSE(3)/πuοΏ½. Note that the stabilizer can also be written as
πuοΏ½ = {(0, exp(π π0Γ))βΆ π β β} = {exp(π (0, π0Γ))βΆ π β β}.
Moreover,π β π°π’(3)fulο¬llsππ = 0if and only ifπ = π‘(0, π0Γ)for someπ‘ β β.
In [DF11], a right inverse toπuοΏ½, i.e. and embedding ofβ3Γ π2intoSE(3), is constructed using Euler angles onSO(3) and spherical coordinates on π2. Unfortunately, this embedding is ill-deο¬ned at the coordinate poles. More generally, there is no such (continuous) embedding: if πuοΏ½βΆ β3 Γ π2 β SE(3) fulο¬lled πuοΏ½ β πuοΏ½(π) = π for all π, restricting it to {0} Γ π2 would yield a continuous map πβΆ π2 β SO(3) such that π(π’) π0 = π’for allπ’ β π2. Now take a vectorπ0 β π0. Thenπ(π’) π0β π’, soπ’ β¦ π(π’) π0 is a non-vanishing, continuous tangential vector ο¬eld onπ2, contradicting the hairy ball theorem2.
An important role in theSE(3)-formalism is played by exponential curves.
Deο¬nition 4.4. Letπ β SE(3)andπ β π°π’(3). The curveπΎ fulο¬lling πΎβ²(π‘) = πΎ(π‘)π, πΎ(0) = π
is calledexponential curvethroughπwith parameterπ.
Using the matrix representation ofπ°π’(3), exponential curves are simply given using the matrix exponential as
πΎ(π‘) = π exp(π‘π). (4.2)
The simplest way to construct exponential curves on the quotient is as follows: given (π₯, π’) β β3Γπ2andπ β π°π’(3), takeπ β πuοΏ½β1(π₯, π’), construct the exponential curve (4.2) and project it back toβ3Γ π2usingπuοΏ½. The question is: under what circumstances does the resulting curve not depend on the arbitrary choice ofπ andπ? For this to hold, we must have
πβ exp(π‘π)π = π exp(π‘π)π for allπ‘ β βandβ β πuοΏ½. Writing this as
exp(βπ‘π) exp(π (0, π0Γ)) exp(π‘π)π = π
forπ β β, diο¬erentiating byπ andπ‘and settingπ = π‘ = 0shows that [(0, π0Γ), π]π = 0.
2The hairy ball theorem states that every continuous, tangential vector ο¬eld onπ2has a zero.
4.1 Choice of the Regularization Functional Withπ = (π, πΓ), this can be expanded as
((π0Γ π) Γ π0 π0Γ π
0 0 ) = 0,
soπ, π β span{π0}. Now
πuοΏ½(π exp((π‘π0, π π0Γ))) = πuοΏ½(π exp(π‘(π0, 0))) = (π₯ + π‘π’, π’) (4.3) is thehorizontal line, which is therefore the only well-deο¬ned exponential curve on the quotient.
On the other hand, the construction above may have been too restrictive, sinceπ β π°π’(3) andπ β πuοΏ½β1(π₯, π’)were chosen ο¬xed and independent of each other. A possible way to improve the situation is to choose the parameterπof the exponential curve depending on the choice ofπ, i.e. to consider
β β π‘ β¦ π exp(π‘π(π))π (4.4)
forπ βΆ πuοΏ½β1(π₯, π’) β π°π’(3). For this to be well-deο¬ned, we need πβ exp(π‘π(πβ))π = π exp(π‘π(π))π
βΊ β exp(π‘π(πβ))ββ1π = exp(π‘π(π))π (4.5) for allβ β πuοΏ½. Diο¬erentiating byπ‘and settingπ‘ = 0yields
Adβπ(πβ)π = π(π)π, (4.6)
whereAdβπ βΆ= βπββ1is theadjoint representation. Explicitly,Adβ(π, πΓ) = (π π, (π π)Γ) forβ = (0, π ). In particular,Adβ(0, π0Γ) = (0, π0Γ)for allβ β πuοΏ½.
Deο¬ne an inner product onπ°π’(3)by
β¨(π1, πΓ1), (π2, πΓ2)β© βΆ= πuοΏ½1π2+ πuοΏ½1π2.
ThenAdβleavesspan{(0, π0Γ)}and its orthogonal complement invariant. π(π)can be uniquely decomposed asπ(π) = πβ(π) + π(π)(0, π0Γ)for someπβΆ πuοΏ½β1(π₯, π’) β βand πβ(π) β (0, π0Γ). Then equation (4.6) implies that
Adβπβ(πβ) = πβ(π). (4.7)
Putting this back into (4.5), we obtain the requirement
exp(π‘πβ(π) + π‘π(πβ)(0, π0Γ))π = exp(π‘πβ(π) + π‘π(π)(0, π0Γ))π.
Diο¬erentiating twice with respect toπ‘, settingπ‘ = 0and using(0, π0Γ)π = 0leads to (π(πβ) β π(π))(0, π0Γ)πβ(π)π = 0.
Writeπβ asπβ(π) = (π(π), π(π)Γ), withπ(π) β π0 by construction, and assume ο¬rst can be summarized by requiring that, without loss of generality,πhas to fulο¬ll
Adβπ(πβ) = π(π) for allβ β πuοΏ½ (4.8) in order for (4.4) to be well-deο¬ned.
Condition (4.8) has a nice interpretation. First, note thatπuοΏ½β1(π₯, π’) β π β¦ ππ(π)is a vector ο¬eld tangential toSE(3). Composing this with the diο¬erential of the projection πuοΏ½ yields a functionπ β¦ ππ(π)π β π(uοΏ½,uοΏ½)(β3Γ π2). Since
πβπ(πβ)π = ππ(π)βπ = ππ(π)π,
this function is in fact constant. Therefore, everyπ βΆ πuοΏ½β1(π₯, π’) β π°π’(3)fulο¬lling (4.8) determines a tangential vector π(uοΏ½,uοΏ½)β π β π(uοΏ½,uοΏ½)(β3 Γ π2). π(uοΏ½,uοΏ½)β is surjective and uοΏ½(π(uοΏ½,uοΏ½)β ) = span{π β¦ (0, π0Γ)}. Thus, the space of functions fulο¬lling (4.8) can be interpreted as the tangent spaceπ(uοΏ½,uοΏ½)(β3Γ π2)plus one additional dimension.
An inverse toπ(uοΏ½,uοΏ½)β can be given explicitly: ifπ(uοΏ½,uοΏ½)β π = (πuοΏ½, πuοΏ½), then
otherwise. The spatial part is a helix with axisπ, while the orientational part performs the corresponding rotation that keeps its components constant with respect to the moving frame of the spatial part. In particular, the curves actually depend onπ β β, so the space of exponential curves at (π₯, π’) is larger than the tangential space π(uοΏ½,uοΏ½)(β3Γ π2).
4.1 Choice of the Regularization Functional In [Fra08], orientation estimation inSE(2)is done by introducingleft-invariant deriva-tives, which are essentially derivatives along exponential curves. Since these curves are parametrized by elements inπ°π’(2), the left-invariant gradientβuοΏ½πof a function πβΆ SE(2) β βcan be interpreted as a map fromSE(2)to the ο¬xed spaceπ°π’(2)β. There-fore, it is possible to take second derivatives, which is in general not possible for the usual derivative. In particular, it is possible to compute theleft-invariant Hessian matrix.
The local orientation is then taken as the (right) singular vector corresponding to the smallest singular value.3 This is motivated by the requirement thatβuοΏ½πshould change as little as possible along the exponential curve corresponding to the local orientation π(π)atπ β SE(2):
π(π) = argmin
β₯uοΏ½β₯=1 β₯πuοΏ½βuοΏ½π(ππΎuοΏ½(π‘))β£uοΏ½=0β₯. (4.9) When applying this to our case, several problems arise:
β’ There are too many exponential curves: they are parametrized by a six-dimen-sional space, while ODFs are only deο¬ned onβ3Γ π2. Intuitively, it seems clear that local knowledge of an ODF can not determine a unique exponential curve.
Indeed, the left-invariant Hessian in this case is singular withleftsingular vector π0Γ. The interpretation of the corresponding rightsingular vector is not clear.
Moreover, numerical experiments showed frequent cases in which there is more than one small singular value; in these cases it is not clear which singular value is the βtrivialβ one and should be discarded.
β’ On the other hand, restricting the space of allowed curves in the most obvious way β namely putting π = 0 in the deο¬nition of π above β does not seem reasonable: it restricts the spatial partπΎuοΏ½ of the curve to be a helix with axis perpendicular to the orientational partπ’of the starting point(π₯, π’) β β3Γ π2. In addition to being a rather arbitrary choice, it also has the undesirable eο¬ect that it is impossible for all points in{π₯} Γ π β β3Γ π2withπ β π2an arbitrarily small, open set to belong to the same oriented structure except for a straight line, π = 0.
β’ A more sensible approach might be to restrict some intrinsic parameters of the projected exponential curve. For example, one might assume that the spatial part of the curve is torsion-free. The torsion ofπΎuοΏ½ can be explicitly calculated to beπ = π β πuοΏ½, leading to an additional indeο¬nite quadratic constraint in the SE(3)-version of (4.9), which is therefore more complicated and time-intensive to solve, and may not even have a unique solution.
β’ In order to estimate local orientation, a (small) minimization problem like (4.9) has to be solved at each point of a discretization ofβ3 Γ π2. While this may be acceptable when solving a diο¬usion equation as a post-processing step, for example by computing one low-dimensional SVD at each point for each time
3Using singular vectors instead of eigenvectors is necessary since the Hessian is in general
3Using singular vectors instead of eigenvectors is necessary since the Hessian is in general