Skip to main content

Matrix Exponential

We seek a solution of a homogeneous first order linear system of differential equations. For illustration purposes we consider the [Graphics:Images/MatrixExponentialMod_gr_1.gif] case:

[Graphics:Images/MatrixExponentialMod_gr_2.gif]
[Graphics:Images/MatrixExponentialMod_gr_3.gif]

First, write the system in vector and matrix form [Graphics:Images/MatrixExponentialMod_gr_4.gif]

[Graphics:Images/MatrixExponentialMod_gr_5.gif].

Then, find the eigenvalues and eigenvectors of the matrix [Graphics:Images/MatrixExponentialMod_gr_6.gif], denote the eigenpairs of A by

[Graphics:Images/MatrixExponentialMod_gr_7.gif] and [Graphics:Images/MatrixExponentialMod_gr_8.gif].

Assumption. Assume that there are two linearly independent eigenvectors [Graphics:Images/MatrixExponentialMod_gr_9.gif], which correspond to the eigenvalues [Graphics:Images/MatrixExponentialMod_gr_10.gif], respectively. Then two linearly independent solution to [Graphics:Images/MatrixExponentialMod_gr_11.gif] are

[Graphics:Images/MatrixExponentialMod_gr_12.gif], and
[Graphics:Images/MatrixExponentialMod_gr_13.gif].

Definition (Fundamental Matrix Solution) The fundamental matrix solution [Graphics:Images/MatrixExponentialMod_gr_14.gif], is formed by using the two column vectors [Graphics:Images/MatrixExponentialMod_gr_15.gif].

(1) [Graphics:Images/MatrixExponentialMod_gr_16.gif].

The general solution to [Graphics:Images/MatrixExponentialMod_gr_17.gif] is the linear combination

(2) [Graphics:Images/MatrixExponentialMod_gr_18.gif].

It can be written in matrix form using the fundamental matrix solution [Graphics:Images/MatrixExponentialMod_gr_19.gif] as follows

[Graphics:Images/MatrixExponentialMod_gr_20.gif].

Notation. When we introduce the notation

[Graphics:Images/MatrixExponentialMod_gr_21.gif],
and
[Graphics:Images/MatrixExponentialMod_gr_22.gif]

The fundamental matrix solution [Graphics:Images/MatrixExponentialMod_gr_23.gif] can be written as

(3) [Graphics:Images/MatrixExponentialMod_gr_24.gif].
or
(4) [Graphics:Images/MatrixExponentialMod_gr_25.gif].

The initial condition [Graphics:Images/MatrixExponentialMod_gr_26.gif]

If we desire to have the initial condition [Graphics:Images/MatrixExponentialMod_gr_27.gif], then this produces the equation

[Graphics:Images/MatrixExponentialMod_gr_28.gif].

The vector of constant [Graphics:Images/MatrixExponentialMod_gr_29.gif] can be solved as follows

[Graphics:Images/MatrixExponentialMod_gr_30.gif].

The solution with the prescribed initial conditions is

[Graphics:Images/MatrixExponentialMod_gr_31.gif].

Observe that [Graphics:Images/MatrixExponentialMod_gr_32.gif] where [Graphics:Images/MatrixExponentialMod_gr_33.gif] is the identity matrix. This leads us to make the following important definition

Definition (Matrix Exponential) If [Graphics:Images/MatrixExponentialMod_gr_34.gif] is a fundamental matrix solution to [Graphics:Images/MatrixExponentialMod_gr_35.gif], then the matrix exponential is defined to be

[Graphics:Images/MatrixExponentialMod_gr_36.gif].

Notation. This can be written as

(5) [Graphics:Images/MatrixExponentialMod_gr_37.gif],
or
(6) [Graphics:Images/MatrixExponentialMod_gr_38.gif].

Fact. For a [Graphics:Images/MatrixExponentialMod_gr_39.gif] system, the initial condition is

[Graphics:Images/MatrixExponentialMod_gr_40.gif],

and the solution with the initial condition [Graphics:Images/MatrixExponentialMod_gr_41.gif] is

[Graphics:Images/MatrixExponentialMod_gr_42.gif],
or
[Graphics:Images/MatrixExponentialMod_gr_43.gif].

Theorem (Matrix Diagonalization) The eigen decomposition of a [Graphics:Images/MatrixExponentialMod_gr_44.gif] square matrix A is

[Graphics:Images/MatrixExponentialMod_gr_45.gif],

which exists when A has a full set of eigenpairs [Graphics:Images/MatrixExponentialMod_gr_46.gif] for [Graphics:Images/MatrixExponentialMod_gr_47.gif], and d is the diagonal matrix

[Graphics:Images/MatrixExponentialMod_gr_48.gif]
and
[Graphics:Images/MatrixExponentialMod_gr_49.gif]

is the augmented matrix whose columns are the eigenvectors of A.

[Graphics:Images/MatrixExponentialMod_gr_50.gif].

Comments

Popular Posts

Runge-Kutta-Fehlberg Method

One way to guarantee accuracy in the solution of an I.V.P. is to solve the problem twice using step sizes h and and compare answers at the mesh points corresponding to the larger step size. But this requires a significant amount of computation for the smaller step size and must be repeated if it is determined that the agreement is not good enough. The Runge-Kutta-Fehlberg method (denoted RKF45) is one way to try to resolve this problem. It has a procedure to determine if the proper step size h is being used. At each step, two different approximations for the solution are made and compared. If the two answers are in close agreement, the approximation is accepted. If the two answers do not agree to a specified accuracy, the step size is reduced. If the answers agree to more significant digits than required, the step size is increased. Each Runge-Kutta-Fehlberg step requires the use of the following six values: Then an approximation to the solution of the I.V.P.

Van Der Pol System

The van der Pol equation is an ordinary differential equation that can be derived from the Rayleigh differential equation by differentiating and setting . It is an equation describing self-sustaining oscillations in which energy is fed into small oscillations and removed from large oscillations. This equation arises in the study of circuits containing vacuum tubes and is given by If , the equation reduces to the equation of simple harmonic motion The van der Pol equation is , where is a constant. When the equation reduces to , and has the familiar solution . Usually the term in equation (1) should be regarded as friction or resistance, and this is the case when the coefficient is positive. However, if the coefficient is negative then we have the case of "negative resistance." In the age of "vacuum tube" radios, the " tetrode vacuum tube " (cathode, grid, plate), was used for a power amplifie

Powell's Method

The essence of Powell's method is to add two steps to the process described in the preceding paragraph. The vector represents, in some sense, the average direction moved over the n intermediate steps in an iteration. Thus the point is determined to be the point at which the minimum of the function f occurs along the vector . As before, f is a function of one variable along this vector and the minimization could be accomplished with an application of the golden ratio or Fibonacci searches. Finally, since the vector was such a good direction, it replaces one of the direction vectors for the next iteration. The iteration is then repeated using the new set of direction vectors to generate a sequence of points . In one step of the iteration instead of a zig-zag path the iteration follows a "dog-leg" path. The process is outlined below. Let be an initial guess at the location of the minimum of the function . Let for be the

Newton-Raphson Method

This method is generally used to improve the result obtained by one of the previous methods. Let x o be an approximate root of f(x) = 0 and let x 1 = x o + h be the correct root so that f(x 1 ) = 0. Expanding f(x o + h) by Taylor ’s series, we obtain f(x o ) + hf’(x o ) + h 2 f’’(x o )/2! +-----= 0 Neglecting the second and higher order derivatives, we have f(x o ) + hf’(x o ) = 0 Which gives h = - f(x o )/f’(x o ) A better approximation than x o is therefore given be x 1 , where x 1 = x o – f(x o )/f’(x o ) Successive approximations are given by x 2 , x 3 ,----,x n+1 , where x n+1 = x n – f(x n )/f’(x n ) Which is the Newton-Raphson formula. Example: - Find a real root of the equation x 3 -5x + 3 = 0. Sol n : - Let, f(x) = x 3 -5x + 3 = 0 f’(x) = 3x 2 - 5 Choosing x o = 1 Step-1: f(x o ) = -1 f(x o ) = -2 So, x 1 =1 – ½ = 0.5 Step-2: f(x 1 ) = 0.625 f’(x 1 ) = -4.25 x 2 = 0.5 + 0.625/4.25 = 0.647 S