• Non ci sono risultati.

n+1 [sin n+1πij ] is unitary!

N/A
N/A
Protected

Academic year: 2021

Condividi "n+1 [sin n+1πij ] is unitary!"

Copied!
8
0
0

Testo completo

(1)

August 21, 2009: I succeed in proving a thing I have believed: q

2

n+1 [sin n+1 πij ] is unitary!

Consider the Fourier matrix of order 2(n + 1):

F 2(n+1) = 1

p2(n + 1) [ω ij 2(n+1) ] 2(n+1) i,j=0 −1 , ω 2(n+1) = e −i

2(n+1)

= e −i

n+1π

. Note that, if o n = p2/(n + 1), then

F 2(n+1) = 1 2 (C − iS),

c ij = o n cos n+1 ijπ , s ij = o n sin n+1 ijπ , i, j = 0, . . . , 2(n + 1) − 1.

Since S and C are real symmetric matrices, we have I = F 2(n+1) F 2(n+1) = 1

2 (C + iS) 1

2 (C − iS) = 1

4 [(C 2 + S 2 ) + i(SC − CS)], Q = F 2(n+1) 2 = 1

2 (C − iS) 1

2 (C − iS) == 1

4 [(C 2 − S 2 ) − i(CS + SC)], being

Q =

 1

J 1 J

, J n × n counter-identity.

As a consequence C 2 + S 2 = 4I C 2 − S 2 = 4Q

CS = SC = 0 ⇒ S 2 = 2(I − Q) = 2

 0

I −J

0

−J I

 .

Now let S 11 , S 12 , S 22 be the n × n matrices defined by the equality

S =

 0

S 11 S 12

0 S 12 T S 22

 ,

that is,

(S 11 ) rs = o n sin n+1 rsπ , (S 12 ) rs = o n sin r(n+1+s)π n+1 , (S 22 ) rs = o n sin (n+1+r)(n+1+s)π

n+1 , 1 ≤ r, s ≤ n.

Observe that S 11 and S 22 are real symmetric and related by the identity S 22 = JS 11 J; moreover S 12 is persymmetric, i.e. S 12 J = JS 12 T . (Recall that S 11 is the (sine) transform diagonalizing the algebra τ of all polynomials in

X =

 1

1 1

1 . ..

. .. 1

1

(2)

).

Since

S 2 =

 0

S 11 2 + S 12 S 12 T S 11 S 12 + S 12 S 22

0

S 12 T S 11 + S 22 S 12 T S 12 T S 12 + S 22 2

 ,

we obtain four identities which in fact reduce to the following only two:

S 11 2 + S 12 S 12 T = 2I, S 11 S 12 J + S 12 JS 11 = −2I. (1) The sum of them yields 0 = S 11 (S 11 +S 12 J)+S 12 J(S 11 +S 12 J) = (S 11 +S 12 J) 2 , but this can happen only if

S 11 + S 12 J = 0, S 12 = −S 11 J (2) (a real symmetric matrix with all the eigenvalues equal to 0 must be null).

Now we are near the thesis. In fact, by (2) the first identity in (1) becomes 2I = S 11 2 + ( −S 11 J)( −S 11 J) T = 2S 11 2 , and so S 11 2 = I.

Remark. From the equality F 2(n+1) = 1 2 (C − iS) it follows that S = i(F 2(n+1) − F 2(n+1) ) = i(I − Q)F 2(n+1) . So, the sine transform of z n × 1, S 11 z, can be computed via a discrete Fourier transform of order 2(n + 1):

i(I − Q)F 2(n+1)

 0 z 0 0

=

 0

S 11 −S 11 J 0

−JS 11 JS 11 J

 0 z 0 0

=

 0 S 11 z

0

−JS 11 z

 .

 Investigate the four submatrices of C, perhaps they also can be expressed in terms of only one and this one is a transform diagonalizing some algebra of matrices . . .

The matrix

A =

 3 2 1 2



does not satisfy the equation A A = AA , thus there is no unitary matrix diagonalizing A. However, T −1 AT is diagonal for a suitable T :

D −1 AD =

 3 √

√ 2

2 2

 , D =

 √ 2 0

0 1

 ,

 α −β

β α

  3 √

√ 2

2 2

  α β

−β α



=

 1 0 0 4



, α = 1

√ 3 , β = r 2 3 , T = 1

√ 3

 √ 2 0

0 1

  1 √

2

− √

2 1



= 1

√ 3

 √

2 2

− √ 2 1

 .

The condition number of T (in the 2-norm), µ 2 (T ) = kT k 2 kT −1 k 2 , is greater than 1:

T T = 1 3

 4 √

√ 2

2 5



⇒ kT k 2 = pρ(T T ) = √

2,

(3)

T −1 =

√ 3 3 √

2

 √ 1 −2 2 √

2



, (T −1 ) (T −1 ) =

 1

2 0

0 1



⇒ kT −1 k 2 = 1.

So, µ 2 (T ) = √

2. Since kT k ∞ = 2+2 3 , kT −1 k ∞ = 3 2 , we have µ (T ) = 1+ √ 2.

Can a non-unitary matrix T have condition number equal to 1 ?

If yes, then, by the Bauer-Fike theorem, the eigenvalue problem would be optimally conditioned for a class of matrices A larger than normal (the A diag- onalized by T , µ 2 (T ) = 1).

A n × n matrix A is said reducible if there exists I ⊂ N = {1, 2, . . . , n}, I 6= ∅, N, such that a ik = 0 for all i ∈ I, k ∈ N \I. Equivalently, A is reducible if there exists a permutation matrix P such that

P T AP =

  n −i ∗ 0  i



,  k k × k matrices, i 6= 0, n (i = |I|, n − i = |N \I|).

Set

C i = {z ∈ C : |z − a ii | <

n

X

j=1,j 6=i

|a ij |}.

It is well known that the subset ∪ n i=1 C i of C includes all the eigenvalues of A (Gershgorin first theorem).

If A is not reducible then we can say something more:

If A is a irreducible n ×n matrix and C i are the inner parts of the Gershgorin disks, then the set ( ∪ n i=1 C i ) ∪ (∩ n i=1 ∂C i ) includes all the eigenvalues of A.

Proof. If λ is an eigenvalue of A, then P

j a ij x j = λx i , P

j,j6=i a ij x j = (λ − a ii )x i ,

|λ − a ii ||x i | ≤ X

j,j6=i

|a ij ||x j |, ∀ i.

Set I = {j : |x j | = kxk ∞ }. Assume I 6= N and let i ∈ I, k ∈ N \I such that a ik 6= 0. Then

|λ − a ii ||x i | ≤ P

j,j6=i |a ij ||x j |

= P

j∈I,j6=i |a ij ||x j | + |a ik ||x k | + P

j∈N \I,j6=k |a ij ||x j |

< P

j∈I,j6=i |a ij ||x i | + |a ik ||x i | + P

j∈N \I,j6=k |a ij ||x i |

= P

j,j 6=i |a ij ||x i |,

|λ − a ii | < P

j,j6=i |a ij |, i.e. λ ∈ C i .

Assume now I = N , that is all entries of the eigenvector x have the same absolute value. In this case:

|λ − a ii ||x i | ≤ X

j,j6=i

|a ij ||x j | = X

j,j6=i

|a ij ||x i |, ∀ i,

|λ − a ii | ≤ P

j,j6=i |a ij |, ∀ i, therefore either λ ∈ C s for some s or λ ∈ ∂C i ∀ i.

 Use the result obtained to prove that any irreducible weakly diagonal dominant n × n matrix A is non singular

 ρ(A) ≤ kAk ∞ .

(4)

By the Gershgorin first theorem, for any eigenvalue λ of A there exists i such that |λ| = |λ − a ii + a ii | ≤ |λ − a ii | + |a ii | ≤ P

j |a ij | ≤ kAk ∞

 If A is irreducible and P

j |a sj | < kAk ∞ for some s, then ρ(A) < kAk ∞ . Given an eigenvalue λ of A, the Gershgorin first theorem for irreducible matrices implies either ∃ i | |λ| = |λ − a ii + a ii | ≤ |λ − a ii | + |a ii | < P

j |a ij | ≤ kAk ∞ or |λ| = |λ − a ii + a ii | ≤ |λ − a ii | + |a ii | = P

j |a ij |, ∀ i, also for i = s, for which we know that P

j |a sj | < kAk ∞

(Jacobi method is able to solve linear systems Ax = b with A weakly diag- onal dominant because in this case the Jacobi iteration matrix J satisfies the conditions ∃ s | P

j |[J] sj | < kJk ∞ and kJk ∞ = 1, thus, by the result of the Exercise, ρ(J) < 1).

Proof of the existence of the SVD of A ∈ C n×n

A n × n ⇒ ∃ U, σ, V , U, V unitary, σ = diag (σ i ) with σ 1 ≥ σ 2 . . . ≥ σ n such that A = U σV .

Proof. Let v 1 , kv 1 k 2 = 1, be such that kAk 2 = kAv 1 k 2 and set u 1 = Av 1 / kAv 1 k 2 ( ku 1 k 2 = 1 and Av 1 = kAk 2 u 1 ). Let ˜ u i , ˜ v i ∈ C n be such that U = [u 1 |˜u 2 | · · · |˜u n ] and V = [v 1 |˜v 2 | · · · |˜v n ] are unitary. Then

U AV =

 u 1

˜ u 2

· · ·

˜ u n

A[v 1 |˜v 2 | · · · |˜v n ] =

 u 1

˜ u 2

· · ·

˜ u n

[ kAk 2 u 1 |A˜v 2 | · · · |A˜v n ] =

 kAk 2 w 0 A ˆ

 ,

kAk 2 = kU AV k 2 = sup v 6=0 k

 kAk 2 w 0 A ˆ

vk

2

kvk

2

k

 kAk 2 w 0 A ˆ

 kAk 2 w

 k

2

k

 kAk 2 w

k

2

≥ √ kAk

22

+ kwk

22

kAk

22

+ kwk

22

= pkAk 2 2 + kwk 2 2

⇒ w = 0 ⇒

kAk 2 = kU AV k 2 = sup v 6=0 k

 kAk 2 0 0 A ˆ

vk

2

kvk

2

≥ sup ˆ v 6=0 k

 kAk 2 0 0 A ˆ

0 ˆ v

k

2

k

0 ˆ v

 k

2

= k ˆ A k 2

⇒ U AV =

 kAk 2 0 0 A ˆ



with ˆ A such that k ˆ A k 2 ≤ kAk 2 .

The thesis follows if we assume it true for matrices of order n − 1.

On SVD: best rank-r approximation of A.

A n × n, A = UσV = P n

1 σ i u i v i , A r = P r

1 σ i u i v i

min {kA − Bk 2 : rank(B) ≤ r} = kA − A r k 2 = σ r+1

(5)

Proof. Let B be a n × n matrix with complex entries whose rank is no more than r and set L = {v : Bv = 0}. Observe that

kA − Bk 2 = sup

v

k(A − B)vk 2 kvk 2 ≥ sup

v ∈L

kAvk 2 kvk 2 .

Set M = Span {v 1 , v 2 , . . . , v r+1 }. Since dim M + dim L ≥ n + 1, there exists z 6= 0, z ∈ M ∩ L,

kA − Bk 2 ≥ kAzk 2

kzk 2 ≥ σ r+1 (first: z ∈ L; second: z ∈ M ⇒ z = P r+1

1 α i v i ⇒ Az = P r+1

1 α i σ i u i ).

Moreover,

kA − A r k 2 = kU diag (0, . . . , 0, σ r+1 , . . . , σ n )V k 2 = k diag (. . .)k 2 = σ r+1

and rank(A r ) ≤ r.

Remark. We also have:

min {kA − Bk F : rank(B) ≤ r} = kA − A r k F = v u u t

n

X

r+1

σ 2 j

In functional analysis for compact operators . . . (linear banded operators on Hilbert spaces) use * as a definition of singular values, approximate an object with something of finite dimension

On SVD: kernel and image of A.

A n × n, A = UσV , σ 1 ≥ . . . ≥ σ k > 0 = σ k+1 = . . . = σ n ⇒ (1) {x ∈ C n : Ax = 0 } = Span {v k+1 , . . . , v n }

(2) {Ax : x ∈ C n } = Span {u 1 , . . . , u k } (3) rank(A) = k = # {σ i : σ i > 0 }

Proof. (1): Ax = 0 iff σV x = 0 iff S k V k x = 0,

S k =

 σ 1

. ..

σ k

 , V k =

 v 1

· · · v k

 ,

iff V k x = 0 iff x is orthogonal to v 1 , . . . , v k iff x is a linear combination of v k+1 , . . . , v n .

(2):

Ax = [U k ]

 S k 0

0 0

  V k





= U k (S k V k x), U k = [u 1 · · · u k ]

⇒ Ax ∈ Span {u 1 , . . . , u k } ⇒ {Ax : x ∈ C n } ⊂ Span {u 1 , . . . , u k }. Now let us show that for any z ∈ C k there exists x ∈ C n , U k z = Ax:

∃ x | Ax = U k z iff

∃ x | U k S k V k x = U k z iff

∃ x | S k V k x = z iff

∃ x | V k x = S k −1 z.

(6)

Since rank(V k ) = k, the latter system admits solution.

On SVD: exercises



A = 1 81

−65 76 104

76 −206 8

104 8 109

 = U DU ,

U = 1 9

−4 4 7

8 1 4

1 8 −4

 , D =

−3 2

−1

 . Write the SVD of A.

 λ i eigenvalues of A ⇒ σ n ≤ |λ i | ≤ σ 1 .

(Ax = λx, A = U σV ⇒ y σ 2 y = x A Ax = |λ| 2 kxk 2 2 . . .).

On SVD: how to compute the rank of a matrix, Gram-Schmidt vs SVD Let a 1 , a 2 , . . . , a m , . . . be a sequence of non null n × 1 vectors and set A m = [a 1 a 2 · · · a m ], m = 1, 2, . . .. There follows an algorithm which computes ma- trices Q m = [q 1 q 2 · · · q m ], n × m, and R m , upper triangular m × m, such that

(1) A m = Q m R m , m = 1, 2, . . .

(2) {q 1 } ∪ {q k : 2 ≤ k ≤ m, a k ∈ Span {a / 1 , . . . , a k−1 }} is an orthonormal basis of the space Span {a 1 , . . . , a m }

(3) if a k , 2 ≤ k ≤ m is linearly dependent from a 1 , . . . , a k−1 , then the k-row of R m is null and q k can be chosen arbitrarily (for instance, q k = 0 or such that Q m Q m = I)

(4) The rank of A m is the number of non null rows of R m

Set ˆ q 1 = a 1 and q 1 = ˆ q 1 / kˆq 1 k 2 . Then a 1 = kˆq 1 k 2 q 1 , i.e.

h a 1

i = h q 1

i [ kˆq 1 k 2 ].

Set ˆ q 2 = a 2 − r 12 q 1 , r 12 such that q 1 q ˆ 2 = 0 (r 12 = q 1 a 2 ) and, if ˆ q 2 6= 0, q 2 = ˆ q 2 / kˆq 2 k 2 . Then a 2 = r 12 q 1 + kˆq 2 k 2 q 2 , i.e.

 a 1 a 2

 =

 q 1 q 2

 kˆq 1 k 2 r 12

0 kˆq 2 k 2

 .

Else, if ˆ q 2 = 0, or, equivalently, a 2 = r 12 q 1 ∈ Span {a 1 }, we can write

 a 1 a 2

 =

 q 1 q 2

 kˆq 1 k 2 r 12

0 0



, q 2 := ˆ q 2 = 0 or arbitrary.

Assume that the first case occurs. Set ˆ q 3 = a 3 −r 13 q 1 −r 23 q 2 , r 13 , r 23 such that

q 1 q ˆ 3 = q 2 q ˆ 3 = 0 (r 13 = q 1 a 3 , r 23 = q 2 a 3 ) and assume ˆ q 3 = 0, or, equivalently,

(7)

a 3 = r 13 q 1 + r 23 q 2 ∈ Span {a 1 , a 2 }. Then we can write:

 a 1 a 2 a 3

 =

 q 1 q 2 q 3

kˆq 1 k 2 r 12 r 13

0 kˆq 2 k 2 r 23

0 0 0

 , q 3 := ˆ q 3 = 0 or arbitrary.

Set ˆ q 4 = a 4 − r 14 q 1 − r 24 q 2 , r 14 , r 24 such that q 1 q ˆ 4 = q 2 q ˆ 4 = 0 (r 14 = q 1 a 4 , r 24 = q 2 a 4 ) and assume ˆ q 4 = 0, or, equivalently, a 4 = r 14 q 1 + r 24 q 2 ∈

Span {a 1 , a 2 }. Then we can write:

 a 1 a 2 a 3 a 4

 =

 q 1 q 2 q 3 q 4

kˆq 1 k 2 r 12 r 13 r 14

0 kˆq 2 k 2 r 23 r 24

0 0 0 0

0 0 0 0

 ,

q 3 := ˆ q 3 = 0, q 4 := ˆ q 4 = 0 or arbitrary.

Set ˆ q 5 = a 5 − r 15 q 1 − r 25 q 2 , r 15 , r 25 such that q 1 q ˆ 5 = q 2 ˆ q 5 = 0 (r 15 = q 1 a 5 , r 25 = q 2 a 5 ) and assume ˆ q 5 6= 0. Set q 5 = ˆ q 5 / kˆq 5 k 2 . Then a 5 = r 15 q 1 +r 25 q 2 + kˆq 5 k 2 q 5 , i.e.

 a 1 a 2 a 3 a 4 a 5

 =

 q 1 q 2 q 3 q 4 q 5

kˆq 1 k 2 r 12 r 13 r 14 r 15

0 kˆq 2 k 2 r 23 r 24 r 25

0 0 0 0 0

0 0 0 0 0

0 0 0 0 kˆq 5 k 2

 ,

q 3 , q 4 null or arbitrary.

. . .

Remark. Since the calculator uses finite arithmetic, the check if ˆ q k , k ≥ 2, is zero or nonzero must be replaced with something of type: kˆq k k is less than ε or not? Moreover, take into account that even a very little perturbation in one entry of a triangular matrix can change the value of its rank (see the following example). These facts imply that the (Gram-Schmidt) algorithm illustrated above may generate a numeric rank of A m which is different from the rank of A m .

Example. Let R be the n × n upper triangular matrix

R =

1 −1 −1 · · · −1 0 1 −1 · · · −1 .. . . .. 1 ... ...

.. . . .. ... −1 0 · · · · 0 1

 .

The rank of R is n, but if the 0 in the (n, 1) entry is replaced with −2 2−n (which for large n is a very little perturbation), then the rank of R becomes n − 1. The SVD of R predicts this observation. In fact, the singular value σ n −1 of R for n = 5, 10, 15 has more or less the same value, 1.5, whereas the smallest singular value, σ n , seems to tend to zero:

n = 5 : σ 5 ≈ 1

10 , n = 10 : σ 10 ≈ 1

100 , n = 15 : σ 15 ≈ 1

10000 .

(8)

So, by examining the singular values of R we see that even if det(R) = 1 (far from zero) for all n, greater is n, smaller is the distance of R from a singular matrix. (Note that R is not normal, in fact µ 2 (R) = σ 1 /σ n ≈ 30, 2000, 10 5

> 1 = max |λ i |/ min |λ i |).

It is known that small perturbations on the entries of A imply at most small perturbations on U, σ, V , A = U σV (SVD problem is well conditioned).

It follows that the algorithm for the computation of the SVD of A can give

accurate approximations of U, σ, V . Having an accurate approximation of σ we

can evaluate precisely the rank of A; we can even quantify how much A is far

from having a smaller rank. Thus it is preferable to compute the rank of a

matrix via SVD, instead via Gram-Schmidt.

Riferimenti

Documenti correlati

Since the aim of this research is to prove the effectiveness of abrasive recycling with an economical approach, the final result will be the economics of the recycled GMA

of the cases were diagnosed as indolent lymphoma, while the remaining were referable to nodular hyperplasia, either purely lymphoid or complex type. Notably, none of the dogs

This is doing matter since 15 billion years ago, when elementary particles joined to form nucleus, electrons, and hydrogen atoms, which started to form new atoms and molecules,

Preconditioned finite element method: the Poisson problem on the square Consider the Poisson problem on Ω = [0, 1] × [0, 1]... Now let us compute the eigenvalues of A (we already

Such decompositions yield formulas for the inverses of Toeplitz, Toeplitz plus Hankel, and Toeplitz plus Hankel-like matrices useful in order to solve Toeplitz plus Hankel-like

The product of a m × m l.t.T. This result is obtained by using a) two representations of a generic Toeplitz matrix involving circulant and (−1)-circulant matrix algebras and b) the

[r]

Answer: the determinant of the matrix is always 0 so the matrix is not invertible for any k, and the rank is either 1 or 2.. So the function is injective, and