The Jordan Canonical Form Just Doesn’t Compute 4

Posted by Cleve Moler,

Camille Jordan (1838-1922)

Photo credit:

(This column first appeared in the May 1994 issue of the MathWorks Newsletter.)

Matrices and differential equations are the key mathematical tools in MATLAB® and Simulink®. The Jordan Canonical Form is the key relationship between matrices and differential equations. So, why doesn't MATLAB use the JCF in any of its computations? In fact, until the Symbolic Math Toolbox came along, we didn't even have a function to compute the JCF.

The difficulty with the Jordan Canonical Form is that it is wildly sensitive to perturbation. Any kind of error -- uncertainty in experimental data, arithmetic roundoff error, linearization of nonlinear functions -- completely changes the JCF and, more importantly, the transformations that generate it.

Let's start with a system of $n$ linear, constant coefficient ordinary differential equations.

$$\dot{x} = Ax$$

Here $x(t)$ is a vector-valued function of $t$, the dot denotes differentiation with respect to $t$, and $A$ is a given $n$ -by- $n$ matrix which is independent of $t$. If you've never heard of -- or want to forget -- the Jordan Canonical Form, then you have to hope there is a matrix $V$ which diagonalizes $A$, that is $V^{-1} A V$ is a diagonal matrix $\Lambda$. Then the change of variables,

$$x = Vy$$

reduces the $n$ -by- $n$ problem to $n$ instances of the 1-by-1 case,

$$\dot{y}_k = \lambda_k y_k$$

The elements of the diagonal matrix $\Lambda$ are the eigenvalues of $A$ and the columns of $V$ are the eigenvectors of $A$. The $k$-th component of the solution $y(t)$ is the exponential function of $t$ determined by the eigenvalue $\lambda_k$, and the initial value $y_k(0)$.

$$ y_k(t) = y_k(0) e^{\lambda_k t} $$

The components of the solution $x(t)$ of the original problem are then linear combinations of the exponential functions determined by the eigenvalues and eigenvectors and the initial values.

But this fantasy about diagonalizing matrices leaves out some very important differential equations, like the single, second order

$$\ddot{u} = 0$$

The solution, of course, is a straight line -- a linear polynomial in $t$. The solution does not involve any exponentials. This equation can be converted to a 2-by-2, first-order system by taking $x$ to be the vector with components

$$x_1 = u$$

$$x_2 = \dot{u}$$

Then the differential equation is

$$\dot{x_1} = x_2$$

$$\dot{x_2} = 0$$

This can be written in matrix form

$$\dot{x} = J x$$

where $J$ is the 2-by-2 matrix generated by the MATLAB statement

  J = [ 0 1; 0 0 ]
  J =
      0     1
      0     0

This matrix cannot be diagonalized; no change of variable which preserves the differential equation will produce a zero in the upper right hand corner. That had better be the case, because the differential equation has polynomial, not exponential, solutions. The matrix J is the simplest example of a nondiagonal Jordan Canonical Form. The eigenvalues -- in this case, a pair of zeros -­ are on the diagonal, but that crucial 1 off the diagonal couples the two differential equations.

A different change of variable leads to a matrix whose structure is not quite so obvious. Let

$$x_1 = u + \dot{u}$$

$$x_2 = u - \dot{u}$$

Then the equation becomes

$$\dot{x} = Ax$$

where $A$ is generated by the MATLAB statement

  A = [ 1/2 -1/2; 1/2 -1/2 ]
  A =
      0.5000   -0.5000
      0.5000   -0.5000

This new change of variables really hasn't altered things very much. The Jordan Canonical Form of A is the matrix J from our first system. The differential equation still has solutions which are linear polynomials in $t$; it can't be transformed into a pair of decoupled equations with exponential solutions.

But now, let's perturb the matrix A very slightly to

  A = [ 0.5 -0.49999999; 0.49999999 -0.5 ]
  A =
    0.500000000000000  -0.499999990000000
    0.499999990000000  -0.500000000000000

The eigenvalues are changed from a pair of zeros to

  lambda = eig(A)
  lambda =
    1.0e-04 *

A change of $10^{-8}$ in the matrix elements changes the eigenvalues by $10^{-4}$. Now this perturbed matrix has distinct eigenvalues, and can be diagonalized. The eigenvector matrix is

[V,~] = eig(A)
V =
   0.707177488329075   0.707036066972953
   0.707036066972953   0.707177488329075

This eigenvector matrix V defines a change of variables that transforms the system into a decoupled system whose solutions involve exp(t/10000) and exp(-t/10000). But we know the solution should be very close to a linear function of $t$. It is probably a lousy idea to try to represent the solution in terms of these two exponentials.

This is reflected in the fact that V is badly conditioned.

   cond(V) = 1.0000e+04

This matrix A is an example of why we can't use the Jordan Canonical Form for practical computation. Technically, A has a full set of linearly independent eigenvectors and a diagonal JCF. But any attempt to use this fact will be plagued by the condition number of V and a resulting loss of accuracy. On the other hand, if we try to say that A does not have a full set of eigenvectors and use a nondiagonal JCF, we are also making possibly significant errors.

Another example can be generated starting with MATLAB's test matrix,

A = gallery(5)
A =
        -9        11       -21        63      -252
        70       -69       141      -421      1684
      -575       575     -1149      3451    -13801
      3891     -3891      7782    -23345     93365
      1024     -1024      2048     -6144     24572

This matrix was constructed so that its characteristic polynomial is

$$\lambda^5 = 0$$

All five eigenvalues are zero. The Jordan Canonical Form can be obtained from the Symbolic Math Toolbox which, in this case, does exact integer calculations with no round off error.

J = jordan(A)
J =
     0     1     0     0     0
     0     0     1     0     0
     0     0     0     1     0
     0     0     0     0     1
     0     0     0     0     0

With a well-conditioned change of variables, the equation

$$ \dot{x} = Ax$$

essentially becomes

$$ \frac{d^4u}{dt^4} = 0 $$

All solutions are cubic polynomials in $t$. So far, so good.

But now change the second diagonal element of A from -69 to

A(2,2) = -70
A =
        -9        11       -21        63      -252
        70       -70       141      -421      1684
      -575       575     -1149      3451    -13801
      3891     -3891      7782    -23345     93365
      1024     -1024      2048     -6144     24572

I want to regard that as a significant change; I am not willing to approximate this new matrix by the unaltered gallery(5). The characteristic polynomial is now


$$(\lambda-1)(\lambda^4 + 2\lambda^3 - 67\lambda^2 + 234\lambda - 168)$$

One of the eigenvalues is $1$. The other four are the distinct, irrational numbers which are the roots of an irreducible quartic. The Jordan Canonical Form is diagonal. We could ask the Symbolic Math Toolbox to compute it, but the result is not very usable.

Instead we can simply ask for

  [V, lambda] = eig(A);
  lambda = diag(lambda)
  lambda =
   -10.5726 + 0.0000i
     3.7953 + 1.3331i
     3.7953 - 1.3331i
     1.0000 + 0.0000i
     0.9820 + 0.0000i

The form of the solution to the differential equation has changed dramatically, from a cubic polynomial in $t$, to a function involving terms like $\exp(-10.5726 t)$ and $\exp(3.7953 t) \sin(1.3331 t)$. Moreover, the change of variables required to obtain this representation is badly conditioned.

  cond(V) = 2.3890e+04

The fundamental difficulty is with matrices, like this last A, which are close to, but not exactly equal to, matrices with non­diagonal JCFs. If we try to diagonalize such matrices, then errors are magnified by the condition number of the eigenvector matrix, which can be arbitrarily large. On the other hand , to use the "nearby" nondiagonal JCF may also represent an unacceptably large error. We're damned if we do and damned if we don't.

The numerically reliable approach to all this is to avoid the Jordan Canonical Form altogether. The portions of MATLAB and Simulink that deal with matrix differential equations use other solution methods, including the Schur Form (which is triangular, instead of bidiagonal) and the matrix exponential (which is not computed from eigenvalues and eigenvectors).

I can't resist finishing up with a personal story. Prof. G. W. "Pete" Stewart of the University of Maryland introduces speakers at conferences by composing original limericks for them. Several years ago, he honored me with:

  Said Professor Cleve Moler one day,
  I'd like to get e to the A.
     It's as easy as heck,
     To get e to the x,
  But I can't get e to the J.

Thanks, Pete.

Get the MATLAB code

Published with MATLAB® R2017a

4 CommentsOldest to Newest

Tahar Loulou replied on : 1 of 4

Dear Professor Moler,

The last year you have published the following post :
What is the Condition Number of a Matrix?
Posted by Cleve Moler, July 17, 2017

In many books on numerical analysis, there is a famous matrix that is presented all the time
when the authors deal with the cond number of matrix. They use the example due to T. S. Wilson where
the matrix A = [5 7 6 5; 7 10 8 7; 6 8 10 9; 5 7, 9 10], x = [1; 1; 1; 1]; and b = [32 23 33 31]. In Matlab language :
A.x = b.
Please could you give us more information on this author and the origin of this matrix.
Any bibliographic reference about this author is welcome.

Thank you very much.
Thank you aslo for Matlab software, it genious tool.

Thank you for the comment, Tahar.
The matrix looks familiar, but I can’t remember where I saw it. I have never used it myself.
Its condition is only 2.9841e+03.
I’m afraid I have no idea who T. S. Wilson is, or was. I don’t recognize that name at all.
— Cleve

Steve Lord replied on : 3 of 4

Tahar Loulou,

A quick Google search found a paper titled “APPLICATION OF THE CONCEPT OF THE CONDITIONING OF A MATRIX TO AN IDENTIFICATION PROBLEM” by Jacques Delforge in “Comp. and Maths with Appls. Vol. 4 pp. 139-148” that has a matrix very similar to the one you posted (the first and second columns were interchanged.) The paper says this example was due to T. S. Wilson and quoted by Bodewig and Todd.

It appears in the form you quoted on page 134 of Bodewig’s “Matrix Calculus” and with the interchanged columns as an exercise on page 48 of Todd’s “Basic Numerical Mathematics: Vol 2: Numerical Algebra” according to Google Books.

Thanks very much, Steve. As I’ve said in this blog before, John Todd was the Caltech professor who got me started in this business. But that still doesn’t tell us who T. S. Wilson was.
— Cleve

Add A Comment

Your email address will not be published. Required fields are marked *

Preview: hide