Skip to main content

QR method

Suppose that A is a real symmetric matrix. Householder’s method is used to construct a similar tridiagonal matrix. Then the QR method is used to find all eigenvalues of the tridiagonal matrix. In the latter construction, plane rotations similar to those that were introduced in Jacobi’s method are used to construct the orthogonal matrices [Graphics:Images/QRmethodMod_gr_1.gif]. The important step the QR method is the factorization [Graphics:Images/QRmethodMod_gr_2.gif] and iteration [Graphics:Images/QRmethodMod_gr_3.gif].

Definition (QR Decomposition): For a nonsingular square matrix [Graphics:Images/QRmethodMod_gr_4.gif], there exists a factorization

[Graphics:Images/QRmethodMod_gr_5.gif]

where [Graphics:Images/QRmethodMod_gr_6.gif] is a unitary and [Graphics:Images/QRmethodMod_gr_7.gif] is an upper triangular matrix.

Remark: [Graphics:Images/QRmethodMod_gr_8.gif] is a unitary means [Graphics:Images/QRmethodMod_gr_9.gif].

For a real nonsingular square matrix [Graphics:Images/QRmethodMod_gr_10.gif], there exists a factorization

[Graphics:Images/QRmethodMod_gr_11.gif]

where [Graphics:Images/QRmethodMod_gr_12.gif] is an orthogonal matrix and [Graphics:Images/QRmethodMod_gr_13.gif] is an upper triangular matrix.

Remark: [Graphics:Images/QRmethodMod_gr_14.gif] is a orthogonal means [Graphics:Images/QRmethodMod_gr_15.gif] (also [Graphics:Images/QRmethodMod_gr_16.gif]).

QR transformation

After finding the [Graphics:Images/QRmethodMod_gr_84.gif] factorization, the [Graphics:Images/QRmethodMod_gr_85.gif] transformation is

[Graphics:Images/QRmethodMod_gr_86.gif].


We now investigate a well known and efficient method for finding all the eigenvalues of a general [Graphics:Images/QRmethodMod_gr_88.gif] real matrix. The [Graphics:Images/QRmethodMod_gr_89.gif] method can be used for an arbitrary real matrix, but for a general matrix it takes many iterations and becomes time consuming.

The [Graphics:Images/QRmethodMod_gr_90.gif] method works much faster on special matrices, preferably:
(i) symmetric tri-diagonal,
(ii) Hessenberg matrices,
(iii) symmetric band matrices.

For this module, we will illustrate the QR method for [Graphics:Images/QRmethodMod_gr_91.gif] real symmetric matrices.

When solving for eigenvalues of a dense symmetric matrix [Graphics:Images/QRmethodMod_gr_92.gif], the standard practice is to reduce the dense matrix to a tridiagonal matrix [Graphics:Images/QRmethodMod_gr_93.gif] through a series of orthogonal transformations, and then to apply an eigenvalue solver for tridiagonal matrices to [Graphics:Images/QRmethodMod_gr_94.gif]. The transformations applied to [Graphics:Images/QRmethodMod_gr_95.gif] preserve eigenvalues, and the eigenvalues of [Graphics:Images/QRmethodMod_gr_96.gif] and [Graphics:Images/QRmethodMod_gr_97.gif] are the same.

The popular eigenvalue solver for symmetric tridiagonal matrices is called the implicit[Graphics:Images/QRmethodMod_gr_98.gif] method. It applies a series of orthogonal transformations [Graphics:Images/QRmethodMod_gr_99.gif] to a tridiagonal matrix [Graphics:Images/QRmethodMod_gr_100.gif] which converges to a diagonal matrix [Graphics:Images/QRmethodMod_gr_101.gif]. Furthermore, [Graphics:Images/QRmethodMod_gr_102.gif] has the same eigenvalues as [Graphics:Images/QRmethodMod_gr_103.gif] which are the diagonal elements of [Graphics:Images/QRmethodMod_gr_104.gif]. In addition, the product of the orthogonal transformations [Graphics:Images/QRmethodMod_gr_105.gif] is a matrix whose columns are the eigenvectors of [Graphics:Images/QRmethodMod_gr_106.gif]. The method is called [Graphics:Images/QRmethodMod_gr_107.gif] because in each iteration the [Graphics:Images/QRmethodMod_gr_108.gif] factorization is computed. The LAPACK routine implementing the implicit [Graphics:Images/QRmethodMod_gr_109.gif] algorithm on tridiagonal symmetric matrices is called DSTEQR.

QR Algorithm: The pseudocode for the QR method is:

1. i = 0
2. [Graphics:Images/QRmethodMod_gr_110.gif]
3. repeat
4. Factor [Graphics:Images/QRmethodMod_gr_111.gif]
5. [Graphics:Images/QRmethodMod_gr_112.gif]
6. i = i+1
7. until convergence

Comments

Popular Posts

Runge-Kutta-Fehlberg Method

One way to guarantee accuracy in the solution of an I.V.P. is to solve the problem twice using step sizes h and and compare answers at the mesh points corresponding to the larger step size. But this requires a significant amount of computation for the smaller step size and must be repeated if it is determined that the agreement is not good enough. The Runge-Kutta-Fehlberg method (denoted RKF45) is one way to try to resolve this problem. It has a procedure to determine if the proper step size h is being used. At each step, two different approximations for the solution are made and compared. If the two answers are in close agreement, the approximation is accepted. If the two answers do not agree to a specified accuracy, the step size is reduced. If the answers agree to more significant digits than required, the step size is increased. Each Runge-Kutta-Fehlberg step requires the use of the following six values: Then an approximation to the solution of the I.V.P.

Van Der Pol System

The van der Pol equation is an ordinary differential equation that can be derived from the Rayleigh differential equation by differentiating and setting . It is an equation describing self-sustaining oscillations in which energy is fed into small oscillations and removed from large oscillations. This equation arises in the study of circuits containing vacuum tubes and is given by If , the equation reduces to the equation of simple harmonic motion The van der Pol equation is , where is a constant. When the equation reduces to , and has the familiar solution . Usually the term in equation (1) should be regarded as friction or resistance, and this is the case when the coefficient is positive. However, if the coefficient is negative then we have the case of "negative resistance." In the age of "vacuum tube" radios, the " tetrode vacuum tube " (cathode, grid, plate), was used for a power amplifie

Powell's Method

The essence of Powell's method is to add two steps to the process described in the preceding paragraph. The vector represents, in some sense, the average direction moved over the n intermediate steps in an iteration. Thus the point is determined to be the point at which the minimum of the function f occurs along the vector . As before, f is a function of one variable along this vector and the minimization could be accomplished with an application of the golden ratio or Fibonacci searches. Finally, since the vector was such a good direction, it replaces one of the direction vectors for the next iteration. The iteration is then repeated using the new set of direction vectors to generate a sequence of points . In one step of the iteration instead of a zig-zag path the iteration follows a "dog-leg" path. The process is outlined below. Let be an initial guess at the location of the minimum of the function . Let for be the

Fibonacci Method

An approach for finding the minimum of in a given interval is to evaluate the function many times and search for a local minimum. To reduce the number of function evaluations it is important to have a good strategy for determining where is to be evaluated. Two efficient bracketing methods are the golden ratio and Fibonacci searches. To use either bracketing method for finding the minimum of , a special condition must be met to ensure that there is a proper minimum in the given interval. The function is unimodal on , if there exists a unique number such that is decreasing on , and is increasing on . In the golden ratio search two function evaluations are made at the first iteration and then only one function evaluation is made for each subsequent iteration. The value of remains constant on each subinterval and the search is terminated at the subinterval, provided that or where are the predefined tolerances. The Fibo