A che serve il clustering attorno a 1
Let A be a p.d. matrix and ε, 0 < ε < 1, be fixed.
Denote by λ ε j the eigenvalues of A outside the interval [1 − ε, 1 + ε] and by r ε the number of such eigenvalues. Set S = [1 − ε, 1 + ε] ∪ {λ ε j } and let p q be the polynomial
p q (λ) = Y
λ
εj1 − λ λ ε j
! T q−r
ε((1 − λ)/ε)
T q−r
ε(1/ε) , q ≥ r ε
where T k (x) denotes the chebycev polynomial of degree k. ((b+a−2λ)/(b−a) = (1 − λ)/ε, (b + a)/(b − a) = 1/ε, if a = 1 − ε, b = 1 + ε ). Notice that S is a set containing all the eigenvalues of A, and p q has exactly degree q and p q (0) = 1.
Then one can say that if x q is the q-th vector generated by the CG method when solving Ax = b, then
kx − x q k A ≤ (max
λ∈S |p q (λ)|)kx − x 0 k A . (bound) This bound for kx − x q k A allows a better evaluation of the CG rate of conver- gence with respect to the well known bound
kx − x q k A ≤ 2
p µ 2 (A) − 1 pµ 2 (A) + 1
! q
kx − x 0 k A , µ 2 (A) = max λ(A)
min λ(A) (wkbound) in case it is known that most of (almost all) the eigenvalues of A are in some interval [1 − ε, 1 + ε] where ε is small (almost zero).
If, moreover, the n×n linear system Ax = b can be seen as one of a sequence of increasing order linear systems, with the property that ∀ ε > 0 ∃ k ε , n ε such that for all n > n ε outside [1 − ε, 1 + ε] fall no more than n ε eigenvalues of A, then (bound) allows to prove the superlinear convergence of CG.
(Note that in general CG has a linear rate of convergence, as a consequence of (wkbound)).
Let us prove these assertions, by evaluating max λ∈S |p q (λ)|.
max λ∈S |p q (λ)| = max λ∈[1−ε,1+ε] |p q (λ)|
≤ (max ... Q
λ
εj1 −
λ λ
εj)(max ...
T
q−rε((1−λ)/ε) T
q−rε(1/ε)
)
= (max ... Q
λ
εj1 −
λ λ
εj)
1 T
q−rε(1/ε) . Now first notice that
T q−r
ε1 ε
= T q−r
ε1+ε 1−ε + 1
1+ε 1−ε − 1
!
> 1 2
q 1+ε
1−ε + 1 q 1+ε
1−ε − 1
q−r
ε.
Then denote by ˆ λ ε j those eigenvalues λ ε j satisfying the inequalities λ ε j < 1 − ε, λ ε j < 1
2 (1 + ε) and observe that
max λ∈[1−ε,1+ε] Q
λ
εj1 −
λ λ
εj≤ max ... Q
ˆ λ
εj1 −
λ λ
εj= Q
λ ˆ
εj1+ε λ ˆ
εj− 1
.
1
So, we have
max λ∈S |p q (λ)| ≤ Q
λ ˆ
εj1+ε λ ˆ
εj− 1
2
q
1+ε 1−ε−1 q
1+ε1−ε
+1
! q−r
ε≤ 2
1+ε
min λ(A) − 1 #ˆ λ
εjq
1+ε 1−ε
−1 q
1+ε1−ε
+1
! q−r
ε≈
1+ε
min λ(A) − 1 #ˆ λ
εjε
qε
rε2
q−rε−1,
where in the latter approximation we have used the following Taylor expansion
f (ε) = q 1+ε
1−ε − 1 q 1+ε
1−ε + 1
= ε 2 + ε 2
2 f 00 (0) + . . . .
Dubbio su GStrang
If t 2 < 1 then Ax i = λ i Ax i , λ i > 0, for n linearly independent vectors x i . Thus, if X denotes the matrix whose columns are the x i then AX = AXD, D = diag (λ i ) ⇒ det(A) 6= 0 ⇒ A −1 Ax i = λ i x i ⇒ i λ i sono gli autovalori di A −1 A.
Since (A −1 A) −1 = A −1 A = L −T L −1 A = L −T (L −1 AL −T )L T , then the eigenvalues of L −1 AL −T are 1/λ i > 0.
⇒ L −1 AL −T is p.d. ⇒ A is p.d. ⇒ A −1 A and E −1 AE −T , EE T = A, have the same eigenvalues.
But, why the x i are independent?
Sui metodi iterativi (H, u k ) r k = b − Ax k
H = A T A: x k+1 = x k + kAu r
TkAu
kk