• Keine Ergebnisse gefunden

My solution

N/A
N/A
Protected

Academic year: 2022

Aktie "My solution"

Copied!
18
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

American Mathematical Monthly

Problem 11415 by Finbarr Holland, generalized

Definitions. a) A matrix is called a complex matrix if all its entries are complex numbers. (In particular, any vector in Cn is considered a n×1 complex matrix.)

b) For any complex matrix A, we denote the matrixAT by A.

c) Let U2(C) be the group of all unitary 2×2 complex matrices. In other words, let

U2(C) =

U ∈GL2(C)|U =U−1 .

d) A complex matrix A is called Hermitian if it satisfies A =A.

Problem. Let A1, A2, ..., An be n Hermitian 2×2 complex matrices. Define a function F from the Cartesian product (U2(C))n toR by

F (U1, U2, ..., Un) = det

n

X

k=1

UkAkUk

!

for every (U1, U2, ..., Un)∈(U2(C))n, Show that

min

U∈(U2(C))n

F (U) =

n

X

k=1

σ1(Ak

n

X

k=1

σ2(Ak),

where σ1(Aj) and σ2(Aj) denote the greatest and the least eigenvalue of the matrix Aj,respectively, for every j ∈ {1,2, ..., n}.

Solution by Darij Grinberg.

We start with an important lemma:

Lemma 1 (the Spectral Theorem for 2×2 matrices). Let A be a Hermitian 2×2 complex matrix.

a)We have TrA∈R, detA∈R and (TrA)2−4 detA∈R≥0. b) Let us define two numbersλ(A) and µ(A) by

λ(A) = 1 2

TrA+ q

(TrA)2−4 detA

; µ(A) = 1

2

TrA− q

(TrA)2−4 detA

.

These numbers λ(A) and µ(A) are real and satisfy λ(A)≥µ(A).

c)If the matrixAis positive definite, then λ(A)≥µ(A)>0.If the matrix A is nonnegative definite, thenλ(A)≥µ(A)≥0.

d) We have λ(A) +µ(A) = TrA and λ(A)·µ(A) = detA. In particular, detA=λ(A)·(TrA−λ(A)).

e)The eigenvalues of A(with algebraic multiplicities) are λ(A) andµ(A).

f ) There exists a matrixS(A)∈U2(C) such that

(S(A))·A·S(A) = diag (λ(A), µ(A)).

(2)

g) We have

λ(A) = max

vAv

vv | v ∈C2\ {0}

; µ(A) = min

vAv

vv | v ∈C2\ {0}

.

h) There exists a matrix Se(A)∈U2(C) such that

Se(A)

·A·Se(A) = diag (λ(A), µ(A)) and det

Se(A)

= 1.

Proof of Lemma 1. SinceAis a 2×2 complex matrix, there exist complex numbers a, b, c, dsuch that A=

a b c d

.

SinceAis Hermitian, we haveA=A,so that

a b c d

=A=A =

a b c d

= a c

b d

, and thus

a=a, b =c, c=b, d =d.

Now, a = a yields a ∈ R, and d = d yields d ∈ R, so that TrA = Tr

a b c d

= a

|{z}

R

+ d

|{z}

R

∈R. Also, a

|{z}

R

− d

|{z}

R

∈R.Besides,

detA = detA = detAT = detA since detBT = detB for any square matrix B

= detA

yields detA∈R.Thus, TrA

| {z }

R

!2

−4 detA

| {z }

R

∈R.Moreover,

detA= det

a b c d

=ad−bc yields

(TrA)2−4 detA=

TrA

| {z }

=a+d

2

−4 ad− b

|{z}

=c

c

!

= (a+d)2−4 (ad−cc) = (a+d)2−4ad

| {z }

=(a−d)2

+4cc

=

a−d

| {z }

R

2

| {z }

R≥0,since squares of reals areR≥0

+4 cc

|{z}

=|c|2R≥0

(1)

∈R≥0.

(3)

Thus, Lemma 1 a)is proven.

Lemma 1a)yields (TrA)2−4 detA∈R≥0,and thus q

(TrA)2−4 detA∈R≥0,so that

λ(A) = 1 2

TrA

| {z }

R

+ q

(TrA)2 −4 detA

| {z }

R≥0R

∈R and

µ(A) = 1 2

TrA

| {z }

R

− q

(TrA)2−4 detA

| {z }

R≥0R

∈R.

In other words, the numbers λ(A) and µ(A) are real. Besides, q

(TrA)2−4 detA ∈ R≥0 yields

q

(TrA)2−4 detA≥ − q

(TrA)2−4 detA, so that

λ(A) = 1 2

TrA+ q

(TrA)2−4 detA

| {z }

≥−

(TrA)2−4 detA

≥ 1 2

TrA− q

(TrA)2−4 detA

=µ(A).

Thus, Lemma 1 b) is proven. Besides, λ(A) +µ(A) = 1

2

TrA+ q

(TrA)2−4 detA

+ 1 2

TrA− q

(TrA)2 −4 detA

= TrA and

λ(A)·µ(A) = 1 2

TrA+ q

(TrA)2−4 detA

· 1 2

TrA− q

(TrA)2−4 detA

= 1 4 ·

TrA+ q

(TrA)2−4 detA TrA− q

(TrA)2 −4 detA

= 1 4 ·

(TrA)2− q

(TrA)2−4 detA 2

| {z }

=(TrA)2−4 detA

= 1

4 ·4 detA= detA.

Thus,

detA=λ(A)· µ(A)

| {z }

=TrA−λ(A), since λ(A)+µ(A)=TrA

=λ(A)·(TrA−λ(A)).

Thus, Lemma 1 d) is proven.

(4)

The characteristic polynomial of the matrix A is det (XI2−A) = det

X 0

0 X

a b c d

= det

X−a 0−b 0−c X−d

= det

X−a −b

−c X−d

= (X−a) (X−d)

| {z }

=X2−(a+d)X+ad

−(−b) (−c)

| {z }

=bc

=X2

a+d

| {z }

=TrA

=λ(A)+µ(A)

 X+

ad−bc

| {z }

=detA

=λ(A)·µ(A)

=X2 −(λ(A) +µ(A))X+λ(A)·µ(A) = (X−λ(A)) (X−µ(A)). Hence, λ(A) and µ(A) are the roots of the characteristic polynomial of the matrix A (with multiplicities). But the eigenvalues of A (with algebraic multiplicities) are the roots of the characteristic polynomial of the matrixA (with multiplicities). Hence, the eigenvalues of A (with algebraic multiplicities) are λ(A) and µ(A). Thus, Lemma 1 e) is proven.

f ) We notice that

diag (α1, α2)·diag (β1, β2) = diag (α1β1, α2β2)

for every α1 ∈C, α2 ∈C, β1 ∈C and β2 ∈C, (2) and

diag (α1, α2) = diag (α1, α2) for every α1 ∈C and α2 ∈C and thus

(diag (α1, α2)) = diag (α1, α2)T = diag (α1, α2)T = diag (α1, α2)

for every α1 ∈C and α2 ∈C. (3) Let ρ =

q

(TrA)2−4 detA. Then, ρ ∈ R≥0 (since (TrA)2 −4 detA ∈ R≥0 by Lemma 1 a)).

Also,

λ(A) = 1 2

TrA+ q

(TrA)2−4 detA

| {z }

= 1

2(TrA+ρ) ;

µ(A) = 1 2

TrA− q

(TrA)2−4 detA

| {z }

= 1

2(TrA−ρ), so that

λ(A)−µ(A) = 1

2(TrA+ρ)−1

2(TrA−ρ) =ρ.

We now distinguish between two cases:

Case 1: We haveρ= 0.

(5)

Case 2: We haveρ6= 0.

First, let us consider Case 2. In this case, ρ6= 0.

Theorem 1 e)yields thatλ(A) is an eigenvalue of the matrixA. Thus, there exists a vector eλ ∈ C2 such that eλ 6= 0 and Aeλ = λ(A)eλ. Since eλ ∈ C2, there exist complex numbers fλ and gλ such that eλ =

fλ gλ

. Then, eλ = eλT = fλ

gλ T

= fλ

gλ T

= fλ, gλ

and thus eλeλ = fλ, gλ fλ

gλ

=fλfλ+gλgλ. Also,eλeλ ∈R>0

(since eλ 6= 0), so that p

eλeλ ∈R>0.

Theorem 1 e)yields thatµ(A) is an eigenvalue of the matrix A. Thus, there exists a vector eµ ∈ C2 such that eµ 6= 0 and Aeµ = µ(A)eµ. Since eµ ∈ C2, there exist complex numbers fµ and gµ such that eµ =

fµ

gµ

. Then, eµ = eµT = fµ

gµ T

= fµ

gµ T

= fµ, gµ

and thuseµeµ= fµ, gµ fµ

gµ

=fµfµ+gµgµ. Also, eµeµ∈R>0

(since eµ6= 0), so that p

eµeµ∈R>0. We have

eλ Aeµ

|{z}

=µ(A)eµ

=eλµ(A)eµ=µ(A)·eλeµ and

eλ A

|{z}

=A

eµ= eλA

| {z }

=(Aeλ)

eµ =

 Aeλ

|{z}

=λ(A)eλ

eµ= (λ(A)eλ)

| {z }

=λ(A)·eλ

eµ=λ(A)·eλeµ,

so that

λ(A)·eλeµ=µ(A)·eλeµ. Thus,

0 = λ(A)·eλeµ−µ(A)·eλeµ =

λ(A)−µ(A)

| {z }

=ρ6=0

·eλeµ, so that eλeµ= 0. Since

eλ

|{z}

=(fλ,gλ) eµ

|{z}

= 0

@

fµ

gµ

1 A

= fλ, gλ fµ

gµ

=fλfµ+gλgµ,

this becomes

fλfµ+gλgµ= 0.

Thus,

fµfλ+gµgλ =fλfµ+gλgµ=fλfµ+gλgµ

since fλ =fλ and gλ =gλ

=fλfµ+gλgµ= 0 = 0.

(6)

Define a matrix W ∈U2(C) by W =

fλ fµ gλ gµ

.

Define a matrix S(A)∈U2(C) by

S(A) =W ·diag 1

peλeλ, 1 peµeµ

!

. (4)

Then,

W =WT =

fλ fµ gλ gµ

T

=

fλ fµ gλ gµ

T

=

fλ gλ fµ gµ

and thus

(S(A)) = W ·diag 1

peλeλ, 1 peµeµ

!!

= diag 1

peλeλ, 1 peµeµ

!!

| {z }

=diag 0

@

1 peλeλ,

1 peµeµ

1 Aby (3)

·W

= diag

1 peλeλ

| {z }

= 1

peλeλ

= 1

peλeλ,

since

eλeλR

, 1

peµeµ

| {z }

= 1

peµeµ

= 1

peµeµ,

since

eµeµR

·W

= diag 1

peλeλ, 1 peµeµ

!

·W, (5)

and W·W =

fλ gλ fµ gµ

·

fλ fµ gλ gµ

=

fλfλ+gλgλ fλfµ+gλgµ fµfλ+gµgλ fµfµ+gµgµ

=

eλeλ 0 0 eµeµ

since fλfλ+gλgλ =eλeλ and fµfµ+gµgµ =eµeµ

= diag eλeλ, eµeµ

, (6)

(7)

so that (4) and (5) yield (S(A)) ·S(A) = diag 1

peλeλ, 1 peµeµ

!

· W·W

| {z }

=diag(eλeλ,eµeµ)

by (6)

·diag 1

peλeλ, 1 peµeµ

!

= diag 1 peλeλ

, 1 peµeµ

!

·diag eλeλ, eµeµ

| {z }

=diag 0

@

1 peλeλ

·eλeλ, 1 peµeµ·e

µeµ

1 A by (2)

·diag 1 peλeλ

, 1 peµeµ

!

= diag

 1

peλeλ ·eλeλ

| {z }

=

eλeλ

, 1 peµeµ

·eµeµ

| {z }

=

eµeµ

·diag 1

peλeλ, 1 peµeµ

!

= diag

peλeλ,p eµeµ

·diag 1

peλeλ, 1 peµeµ

!

= diag p

eλeλ· 1 peλeλ,p

eµeµ· 1 peµeµ

!

(by (2))

= diag (1,1) =I2.

Thus, the matrix S(A) is left-invertible. Hence, the matrix S(A) is invertible (since every left-invertible matrix is invertible). In other words, S(A) ∈ GL2(C). Besides,

(S(A)) = (S(A))−1(since (S(A))·S(A) = I2). Thus,S(A)∈ {U ∈GL2(C)|U =U−1}= U2(C).

On the other hand, afλ+bgλ

cfλ+dgλ

=

a b c d

| {z }

=A

fλ gλ

| {z }

=eλ

=Aeλ =λ(A)eλ =λ(A) fλ

gλ

=

λ(A)fλ λ(A)gλ

yields

afλ+bgλ =λ(A)fλ and cfλ+dgλ =λ(A)gλ. (7) Also,

afµ+bgµ cfµ+dgµ

=

a b c d

| {z }

=A

fµ gµ

| {z }

=eµ

=Aeµ=µ(A)eµ=µ(A) fµ

gµ

=

µ(A)fµ µ(A)gµ

yields

afµ+bgµ=µ(A)fµ and cfµ+dgµ=µ(A)gµ. (8)

(8)

Now,

A·W =

a b c d

·

fλ fµ gλ gµ

=

afλ+bgλ afµ+bgµ cfλ +dgλ cfµ+dgµ

=

λ(A)fλ µ(A)fµ λ(A)gλ µ(A)gµ

(by (7) and (8)) and

W

|{z}

= 0

@

fλ fµ

gλ gµ

1 A

·diag (λ(A), µ(A))

| {z }

= 0

@

λ(A) 0 0 µ(A)

1 A

=

fλ fµ gλ gµ

·

λ(A) 0 0 µ(A)

=

fλ·λ(A) +fµ·0 fλ·0 +fµ·µ(A) gλ·λ(A) +gµ·0 gλ·0 +gµ·µ(A)

=

fλ·λ(A) fµ·µ(A) gλ ·λ(A) gµ·µ(A)

=

λ(A)fλ µ(A)fµ λ(A)gλ µ(A)gµ

yieldA·W =W ·diag (λ(A), µ(A)), so that (4) and (5) yield

(S(A))·A·S(A)

= diag 1 peλeλ

, 1 peµeµ

!

·W· A·W

| {z }

=W·diag(λ(A),µ(A))

·diag 1 peλeλ

, 1 peµeµ

!

= diag 1

peλeλ, 1 peµeµ

!

· W·W

| {z }

=diag(eλeλ,eµeµ)

by (6)

·diag (λ(A), µ(A))·diag 1

peλeλ, 1 peµeµ

!

= diag 1 peλeλ

, 1 peµeµ

!

·diag eλeλ, eµeµ

| {z }

=diag 0

@

1 peλeλ

·eλeλ, 1 peµeµ·e

µeµ

1 A by (2)

·diag (λ(A), µ(A))·diag 1 peλeλ

, 1 peµeµ

!

= diag

 1

peλeλ ·eλeλ

| {z }

=

eλeλ

, 1

peµeµ ·eµeµ

| {z }

=

eµeµ

·diag (λ(A), µ(A))·diag 1

peλeλ, 1 peµeµ

!

= diag

peλeλ,p eµeµ

·diag (λ(A), µ(A))

| {z }

=diag(√

eλeλ·λ(A),

eµeµ·µ(A))

by (2)

·diag 1

peλeλ, 1 peµeµ

!

= diag

peλeλ·λ(A),p

eµeµ·µ(A)

·diag 1

peλeλ, 1 peµeµ

!

= diag p

eλeλ·λ(A)· 1 peλeλ

,p

eµeµ·µ(A)· 1 peµeµ

!

(by (2))

= diag (λ(A), µ(A)).

(9)

Thus, we have proven that, in Case 2, there exists a matrix S(A)∈U2(C) such that (S(A))·A·S(A) = diag (λ(A), µ(A)).

In other words, we have proven that Lemma 1 f )holds in Case 2.

Next, let us consider Case 1. In this case, ρ= 0, so that 0 =ρ2 =

q

(TrA)2−4 detA 2

= (TrA)2−4 detA

=

a−d

| {z }

R

2

+ 4 cc

|{z}

=|c|2R

(by (1)) (9)

≥4 cc

|{z}

=|c|2

since (a−d)2 ≥0, since a−d∈R and since squares of reals are ≥0

= 4|c|2,

so that 0≥ |c|2. But |c|2 ≥0 (since |c| ∈ R and since squares of reals are ≥0). Thus,

|c|2 = 0, so that |c| = 0 and thus c = 0. Hence, b = c = 0 = 0. Now, (9) yields 0 = (a−d)2+ 4c c

|{z}

=0

= (a−d)2, so that a−d= 0, so that a=d. Thus,

A =

a b c d

=

d 0 0 d

(since b = 0, c= 0 and a=d)

= diag (d, d),

so that TrA=d+d= 2d. Hence,

λ(A) = 1 2

TrA

| {z }

=2d

+ ρ

|{z}

=0

= 1

2(2d+ 0) =d;

µ(A) = 1 2

TrA

| {z }

=2d

− ρ

|{z}

=0

= 1

2(2d−0) = d, so that diag (λ(A), µ(A)) = diag (d, d).

Now, let S(A) = I2. Then, clearly, S(A)∈U2(C) and (S(A))

| {z }

=I2=I2

·A·S(A)

| {z }

=I2

=I2·A·I2 =A= diag (d, d) = diag (λ(A), µ(A)).

Thus, we have proven that, in Case 1, there exists a matrix S(A)∈U2(C) such that (S(A))·A·S(A) = diag (λ(A), µ(A)).

In other words, we have proven that Lemma 1 f )holds in Case 1.

Altogether, we have now proven that Lemma 1 f ) holds in both Cases 1 and 2.

Thus, Lemma 1 f )always holds. Hence, Lemma 1 f )is proven.

(10)

g) According to Lemma 1 f ), there exists a matrix S(A)∈U2(C) such that (S(A))·A·S(A) = diag (λ(A), µ(A)).

We have S(A) ∈ GL2(C) (since S(A) ∈ U2(C) = {U ∈GL2(C)|U =U−1}). In other words, the matrix S(A) is invertible. Also, S(A) ∈ U2(C) yields (S(A)) = (S(A))−1, so that (S(A))S(A) =I2.

We have

C2\ {0} ⊇

S(A)w | w∈C2\ {0}

(since S(A)w∈C2\ {0}for every w∈C2\ {0} (since the matrixS(A) is invertible)) and

C2\ {0} ⊆

S(A)w | w∈C2\ {0}

(since for every v ∈ C2\ {0}, there exists some w ∈ C2 \ {0} such that v = S(A)w (in fact, let w= (S(A))−1v; then,v =S(A)wand w∈C2\ {0} (since S(A)w=v ∈ C2\ {0}, so that S(A)w6= 0, so thatw6= 0))). Thus,

C2\ {0}=

S(A)w | w∈C2\ {0} . Hence,

vAv

vv | v ∈C2\ {0}

=

vAv

vv | v ∈

S(A)w | w∈C2\ {0}

=

(S(A)w)AS(A)w

(S(A)w)S(A)w | w∈C2 \ {0}

=

w(S(A))AS(A)w

w(S(A))S(A)w | w∈C2\ {0}

(since (S(A)w) =w(S(A)))

=

wdiag (λ(A), µ(A))w

ww | w∈C2\ {0}

(10)

since (S(A))AS(A) = diag (λ(A), µ(A)) andw(S(A))S(A)

| {z }

=I2

w=wI2w=ww

. Now, for every w ∈ C2 \ {0}, there exist w1 ∈ C and w2 ∈ C such that w =

w1 w2

, and we havew =wT =

w1 w2

T

= w1

w2 T

= (w1, w2), so that ww= (w1, w2)

w1 w2

=w1w1

| {z }

=|w1|2

+w2w2

| {z }

=|w2|2

=|w1|2+|w2|2 and

wdiag (λ(A), µ(A))w= (w1, w2) diag (λ(A), µ(A)) w1

w2

| {z }

= 0

@

λ(A)w1 µ(A)w2

1 A

= (w1, w2)

λ(A)w1 µ(A)w2

=w1λ(A)w1+w2µ(A)w2 =λ(A)w1w1

| {z }

=|w1|2

+µ(A)w2w2

| {z }

=|w2|2

=λ(A)|w1|2+µ(A)|w2|2.

(11)

Hence,

wdiag (λ(A), µ(A))w

ww = λ(A)|w1|2+µ(A)|w2|2

|w1|2+|w2|2 (11)

≤ λ(A)|w1|2 +λ(A)|w2|2

|w1|2 +|w2|2

since µ(A)≤λ(A) and since |w2|2 ≥0 and |w1|2+|w2|2 ≥0

= λ(A) |w1|2+|w2|2

|w1|2+|w2|2 =λ(A),

and there exists aw∈C2\ {0}satisfying wdiag (λ(A), µ(A))w

ww =λ(A) (in fact, set w=

1 0

; then,w =wT = 1

0 T

= 1

0 T

= 1

0 T

= (1,0) and thus

wdiag (λ(A), µ(A))w

ww =

(1,0) diag (λ(A), µ(A)) 1

0

(1,0) 1

0

=

(1,0)

λ(A) 0

(1,0) 1

0

since diag (λ(A), µ(A)) 1

0

=

λ(A)·1 µ(A)·0

=

λ(A) 0

= 1·λ(A) + 0·0

1·1 + 0·0 = 1·λ(A)

1·1 =λ(A) ). Thus,

max

wdiag (λ(A), µ(A))w

ww | w∈C2\ {0}

=λ(A). (12) Besides, for every w ∈ C2 \ {0}, there exist w1 ∈ C and w2 ∈ C such that w = w1

w2

, and (11) yields wdiag (λ(A), µ(A))w

ww = λ(A)|w1|2+µ(A)|w2|2

|w1|2+|w2|2 ≥ µ(A)|w1|2+µ(A)|w2|2

|w1|2+|w2|2

since λ(A)≥µ(A) and since |w1|2 ≥0 and |w1|2+|w2|2 ≥0

= µ(A) |w1|2+|w2|2

|w1|2+|w2|2 =µ(A),

and there exists aw∈C2\ {0} satisfying wdiag (λ(A), µ(A))w

ww =µ(A) (in fact, set

(12)

w= 0

1

; then,w =wT = 0

1 T

= 0

1 T

= 0

1 T

= (0,1) and thus

wdiag (λ(A), µ(A))w

ww =

(0,1) diag (λ(A), µ(A)) 0

1

(0,1) 0

1

=

(0,1) 0

µ(A)

(0,1) 0

1

since diag (λ(A), µ(A)) 0

1

=

λ(A)·0 µ(A)·1

= 0

µ(A)

= 0·0 + 1·µ(A)

0·0 + 1·1 = 1·µ(A)

1·1 =µ(A) ). Thus,

min

wdiag (λ(A), µ(A))w

ww | w∈C2\ {0}

=µ(A). (13)

Now,

λ(A) = max

wdiag (λ(A), µ(A))w

ww | w∈C2\ {0}

(by (12))

= max

vAv

vv | v ∈C2\ {0}

(by (10)) and

µ(A) = min

wdiag (λ(A), µ(A))w

ww | w∈C2\ {0}

(by (13))

= min

vAv

vv | v ∈C2\ {0}

(by (10)). Thus, Lemma 1 g) is proven.

c) According to Lemma 1 f ), there exists a matrix S(A)∈U2(C) such that (S(A))·A·S(A) = diag (λ(A), µ(A)).

We have S(A) ∈ GL2(C) (since S(A) ∈ U2(C) = {U ∈GL2(C)|U =U−1}). In other words, the matrixS(A) is invertible. Thus,S(A)·

0 1

6= 0 (since 0

1

6= 0).

Letv =S(A)· 0

1

. Then,v =S(A)· 0

1

6= 0.

(13)

Now, vAv=

S(A)· 0

1

| {z }

= 0

@

0 1

1 A

·(S(A))

·A·

S(A)· 0

1

= 0

1

·(S(A))·A·S(A)

| {z }

=diag(λ(A),µ(A))

· 0

1

=

0 1

| {z }

= 0

@

0 1

1 A

T

= 0

@

0 1

1 A

T

=(0,1)=(0,1)

·diag (λ(A), µ(A))· 0

1

| {z }

= 0

@

λ(A)·0 µ(A)·1

1 A=

0

@

0 µ(A)

1 A

= (0,1)· 0

µ(A)

= 0·0 + 1·µ(A) = 1·µ(A) = µ(A). (14)

Now, if the matrix A is positive definite, then vAv > 0 (since v 6= 0), what becomes µ(A)> 0 (by (14)), so that λ(A) ≥ µ(A) > 0 (since λ(A) ≥ µ(A)). Besides, if the matrix A is nonnegative definite, then vAv ≥ 0, what becomes µ(A) ≥ 0 (by (14)), so that λ(A)≥µ(A)≥0 (sinceλ(A)≥µ(A)). Thus, Lemma 1 c)is proven.

h) According to Lemma 1 f ), there exists a matrix S(A)∈U2(C) such that (S(A))·A·S(A) = diag (λ(A), µ(A)).

Note that S(A) ∈ U2(C) = {U ∈GL2(C)|U =U−1} yields S(A) ∈ GL2(C) and (S(A)) = (S(A))−1, and thus (S(A))·S(A) =I2. Thus,

1 = det I2

|{z}

=(S(A))·S(A)

= det ((S(A))·S(A)) = det

(S(A))

| {z }

=S(A)T

·det (S(A))

= det

S(A)T

| {z }

=detS(A)

=det(S(A))

·det (S(A)) = det (S(A))·det (S(A)) =|det (S(A))|2,

so that |det (S(A))|= 1 (since |det (S(A))| ∈R≥0).

Since the field C is algebraically closed, there exists some τ ∈ C such that τ2 = det (S(A)). We have |τ|2 =|τ2|=|det (S(A))|= 1, so that |τ|= 1 (since |τ| ∈R≥0), and thus τ 6= 0 and τ τ =|τ|2 = 12 = 1, so that τ = 1

τ. Define a matrix Se(A) ∈ M2(C) by Se(A) = 1

τS(A). Then, Se(A) = 1

τS(A) ∈

(14)

GL2(C) (sinceS(A)∈GL2(C) and 1

τ 6= 0) and

Se(A)

= 1

τS(A)

= 1

τ

|{z}

−1−1

(S(A))−1(S(A))

| {z }

=(S(A))−1

−1(S(A))−1 =

 τ

|{z}

=1 τ

S(A)

−1

= 1

τS(A) −1

=

Se(A)−1 .

Thus,Se(A)∈ {U ∈GL2(C)|U =U−1}= U2(C). Besides,

Se(A)

·A·Se(A) = 1

τS(A)

| {z }

−1(S(A))

·A· 1

τS(A) =τ−1(S(A))·A· 1 τS(A)

= 1

τ τ

|{z}

=1,since τ τ=1

(S(A))·A·S(A)

| {z }

=diag(λ(A),µ(A))

= diag (λ(A), µ(A))

and

det

Se(A)

= det 1

τS(A)

= 1

τ 2

det (S(A))

| {z }

2

= 1

τ 2

τ2 = 1.

Thus, we have proven that there exists a matrix Se(A)∈U2(C) such that

Se(A)

·A·Se(A) = diag (λ(A), µ(A)) and det

Se(A)

= 1.

In other words, we have proven Lemma 1 h).

A conclusion from Lemma 1:

Lemma 2. Let B1, B2, ..., Bn be n Hermitian 2×2 complex matrices.

Then, λ n

P

k=1

Bk

n

P

k=1

λ(Bk) and

det

n

X

k=1

Bk

!

n

X

k=1

λ(Bk

n

X

k=1

µ(Bk).

Proof of Lemma 2. For every k ∈ {1,2, ..., n}, we have λ(Bk) = max

vBkv

vv | v ∈C2\ {0}

(by Lemma 1 g), applied to A=Bk).

(15)

Also,

λ

n

X

k=1

Bk

!

= max





 v

n P

k=1

Bk

v

vv | v ∈C2\ {0}





by Lemma 1 g), applied to A=

n

X

k=1

Bk

!

= max





n

P

k=1

vBkv

vv | v ∈C2\ {0}









n

P

k=1

vBkv

vv | v ∈C2\ {0}





 .

Hence, there exists somew∈C2\ {0} such that λ n

P

k=1

Bk

=

n

P

k=1

wBkw

ww . Thus,

λ

n

X

k=1

Bk

!

=

n

P

k=1

wBkw ww =

n

X

k=1

wBkw ww ≤

n

X

k=1

λ(Bk)

since wBkw ww ∈

vBkv

vv | v ∈C2\ {0}

and thus wBkw

ww ≤max

vBkv

vv | v ∈C2\ {0}

=λ(Bk) for every k∈ {1,2, ..., n}

 .

Now, let Be =

n

P

k=1

Bk. Then,

TrBe = Tr

n

X

k=1

Bk

!

=

n

X

k=1

TrBk (since the trace is linear).

Now, λ Be

+µ Be

= TrBe (by Lemma 1 d), applied to A=Be).

Besides, let λΣ =

n

P

k=1

λ(Bk). Then, λ Be

= λ n

P

k=1

Bk

n

P

k=1

λ(Bk) = λΣ, so that λ

Be

−λΣ ≤0. Also, TrBe−

λ Be

Σ

≤TrBe− λ

Be +λ

Be

(since λΣ ≥v)

≤TrBe−

 λ

Be

Be

| {z }

=TrBe

since λ Be

≥µ Be

by Lemma 1 b), applied to A=Be

= TrBe−TrBe= 0.

(16)

Thus,

λ Be

·

TrBe−λ Be

−λΣ·

TrBe−λΣ

=

λ Be

·TrBe− λ

Be2

λΣ·TrBe−λ2Σ

=

 λ

Be

·TrBe−λΣ·TrBe

| {z }

=(λ(Be)−λΣ)·TrBe

λ Be2

−λ2Σ

| {z }

=(λ(Be)−λΣ)(λ(Be)Σ)

= λ

Be

−λΣ

·TrBe− λ

Be

−λΣ λ Be

Σ

=

 λ

Be

−λΣ

| {z }

≤0

·

TrBe− λ

Be

Σ

| {z }

≤0

≥0,

so that

λ Be

·

TrBe−λ Be

≥λΣ·

TrBe−λΣ

. (15)

Hence, det

n

X

k=1

Bk

!

= detBe =λ Be

·

TrBe−λ

Be

by Lemma 1 d), applied to A=Be

≥λΣ·

TrBe−λΣ

(by (15))

=

n

X

k=1

λ(Bk

n

X

k=1

TrBk

n

X

k=1

λ(Bk)

!

since λΣ =

n

X

k=1

λ(Bk) and TrBe=

n

X

k=1

TrBk

!

=

n

X

k=1

λ(Bk

n

X

k=1

µ(Bk)

since

n

P

k=1

TrBk− Pn

k=1

λ(Bk) =

n

P

k=1

(TrBk−λ(Bk)) =

n

P

k=1

µ(Bk), because TrBk−λ(Bk) =µ(Bk) for every k ∈ {1,2, ..., n}

(because λ(Bk) +µ(Bk) = TrBk by Lemma 1 d), applied to A=Bk)

 ,

and Lemma 2 is proven.

Now let us solve the problem:

For every k ∈ {1,2, ..., n}, there exists a matrix S(Ak)∈U2(C) such that

(S(Ak))·Ak·S(Ak) = diag (λ(Ak), µ(Ak)) (16) (according to Lemma 1 f ), applied to A = Ak). These matrices S(A1), S(A2), ...,

(17)

S(An) satisfy (S(A1), S(A2), ..., S(An))∈(U2(C))n and

F (S(A1), S(A2), ..., S(An)) = det

n

X

k=1

(S(Ak))·Ak·S(Ak)

| {z }

=diag(λ(Ak),µ(Ak)) by (16)

= det

n

X

k=1

diag (λ(Ak), µ(Ak))

| {z }

=diag

n P

k=1

λ(Ak),

n

P

k=1

µ(Ak)

«

= det diag

n

X

k=1

λ(Ak),

n

X

k=1

µ(Ak)

!!

=

n

X

k=1

λ(Ak

n

X

k=1

µ(Ak). Hence,

n

X

k=1

λ(Ak

n

X

k=1

µ(Ak)∈ {F(U) | U ∈(U2(C))n}. (17) On the other hand, for every (U1, U2, ..., Un)∈(U2(C))n, we have

F (U1, U2, ..., Un) = det

n

X

k=1

UkAkUk

!

n

X

k=1

λ(UkAkUk

n

X

k=1

µ(UkAkUk)

by Lemma 2, applied to Bk =UkAkUk, because the matrix UkAkUk is Hermitian for everyk ∈ {1,2, ..., n}, since (UkAkUk) =UkAk(Uk) =UkAkUk,

because Ak =Ak (since the matrix Ak is Hermitian) and (Uk) =Uk

=

n

X

k=1

λ(Ak

n

X

k=1

µ(Ak)

since for every k ∈ {1,2, ..., n}, we have Uk =Uk−1 (since Uk∈U2(C) ) and thus UkAkUk =Uk−1AkUk, so that the matrices UkAkUk and Ak are similar, so that

Tr (UkAkUk) = TrAk and det (UkAkUk) = detAk, so that λ(UkAkUk) = 1

2

Tr (UkAkUk) + q

(Tr (UkAkUk))2−4 det (UkAkUk)

= 1 2

TrAk+ q

(TrAk)2−4 detAk

=λ(Ak) and µ(UkAkUk) = 1

2

Tr (UkAkUk)− q

(Tr (UkAkUk))2−4 det (UkAkUk)

= 1 2

TrAk− q

(TrAk)2 −4 detAk

=µ(Ak)

 .

In other words, for every U ∈(U2(C))n, we have F (U)≥

n

X

k=1

λ(Ak

n

X

k=1

µ(Ak).

(18)

This, together with (17), yields min

U∈(U2(C))nF (U) =

n

X

k=1

λ(Ak

n

X

k=1

µ(Ak). (18) Lemma 1 b) and e) yields that for every Hermitian 2×2 complex matrix A, the numbers λ(A) and µ(A) are the greatest and the least eigenvalue of the matrix A, respectively. Hence, λ(Aj) =σ1(Aj) and µ(Aj) =σ2(Aj) for every j ∈ {1,2, ..., n}. Thus, (18) becomes

min

U∈(U2(C))n

F (U) =

n

X

k=1

σ1(Ak

n

X

k=1

σ2(Ak), and the problem is solved.

Referenzen

ÄHNLICHE DOKUMENTE

Pbtscher (1983) used simultaneous Lagrange multiplier statistics in order to test the parameters of ARMA models; he proved the strong consistency of his procedure

The first subsection shows that the estimate based on the inclusion algorithm, namely the matricial matrix cube algorithm of Subsection 4.5, is essentially identical to that obtained

Rumpf: A level set method for anisotropic geometric diffusion in 3D

The above formula can be understood as a linear expansion of E into contributions coming from first-neighbour interactions (the first sum) and contributions

This multilayered and complex policy involved the control of inflation; the reform of the domestic economy, including the banking system, the independence of the

In our paper “Killip-Simon problem and Jacobi flow on GMP matrices” we generalize this result to the finite system of intervals

20 The aims of this retrospective study were to document the clinical presentation, clinical course, and treatment of the rare entity of primary LMS of the oral cavity and to

The standard way of extending the Square attack to six rounds (in the case of a chosen ciphertext attack) is by guessing four bytes of the whitening key and peeling of the first