Let be distinct real numbers, and consider the matrix
Prove that can be expressed as
where the column vector generates a polynomial
satisfying
You may cite anything you know about polynomial interpolation to justify that is nonsingular.
|
We want to show that
or equivalently the th, th entry of is for and for i.e.
First notice that
Also notice that
Hence,
Let be a real matrix of order with eigenvalues and linearly independent eigenvectors .
Suppose you want to find an eigenvector corresponding to the eigenvalue and you are given such that
Specify a numerical algorithm for finding and give a convergence proof for this algorithm demonstrating that it is convergent under appropriate circumstances
|
Let
Then,
which implies
Since shifting the eigenvalues and inverting the matrix does not affect the eigenvectors,
Assume for all . Generate to find . Start with arbitrary such that .
For
(Rayleigh quotient)
End
If , for all , then will be dominated by .
Since are linearly independent, they form a basis of . Hence,
From the definition of eigenvectors,
To find a general form of , the approximate eigenvector at the kth step, examine a few steps of the algorithm:
From induction,
Hence,
Comparing a weighted and ,
since by assumption.
The above expression goes to as since , for all . Hence as grows, is parallel to . Because , it must be that .
Suppose A is an upper triangular, nonsingular matrix. Show that both Jacobi iterations and Gauss-Seidel iterations converge in finitely many steps when used to solve .
|
Let where is a diagonal matrix, is a lower triangular matrix with a zero diagonal and is an upper triangular matrix also with a zero diagonal.
The Jacobi iteration can be found by substituting into , grouping , and solving for i.e.
Since by hypothesis, the iteration is
Similarly, the Gauss-Seidel iteration can be found by substituting into , grouping , and solving for i.e.
Since by hypothesis, the iteration has identical from as Jacopi:
Jacobi and Gauss-Seidel are iterative methods that split the matrix into , , and : Diagonal, Upper (everything above the diagonal), and Lower (everything below the diagonal) triangular matrices, respectively. Their iterations go as such
- (Gauss-Seidel)
- (Jacobi)
In our case is upper triangular, so is the zero matrix. As a result, the Gauss-Seidel and Jacobi methods take on the following identical form
Additionally, can be written
Subtracting from we get
In our problem, is diagonal and is upper triangular with zeros along the diagonal. Notice that the product will also be upper triangular with zeros along the diagonal.
Let ,
Also, let be the related matrix
Where is , is , and is .
Finally, the product (call it (1)) is
Here is almost identical in structure to , except that its diagonal elements are zeros.
At this point the convergence in steps (the size of the starting matrix) should be apparent since is just where and each time is multiplied by , the k-th super-diagonal is zeroed out (where k=0 is the diagonal itself). After applications of , the result will be the zero matrix of size .
In brief, is the zero matrix of size .
Therefore , i.e. the Jacobi and Gauss-Seidel Methods used to solve converge in steps when is upper triangular.