Significance of Jordan Canonical Form

Click For Summary
SUMMARY

The significance of Jordan Canonical Form (JCF) lies in its ability to handle matrices that cannot be diagonalized, providing a reliable alternative for solving systems of linear differential and recurrence equations. JCF simplifies calculations, such as computing powers of matrices, by transforming them into a nearly diagonal structure, which allows for easier manipulation of matrix operations. The discussion emphasizes that while numerical algorithms exist for matrix calculations, understanding JCF is crucial for obtaining analytic expressions and insights into matrix behavior, particularly in the context of differential equations.

PREREQUISITES
  • Understanding of linear algebra concepts, particularly matrix theory.
  • Familiarity with matrix diagonalization and its limitations.
  • Basic knowledge of differential equations and their matrix representations.
  • Experience with matrix operations, including multiplication and inversion.
NEXT STEPS
  • Study the process of converting matrices to Jordan Canonical Form using specific algorithms.
  • Explore the applications of Jordan Canonical Form in solving linear differential equations.
  • Learn about the relationship between Jordan Canonical Form and singular value decomposition.
  • Investigate the implications of Jordan Canonical Form in the context of linear transformations and coordinate changes.
USEFUL FOR

Mathematicians, engineers, and students working with linear algebra, differential equations, and matrix computations will benefit from this discussion, particularly those seeking to deepen their understanding of matrix theory and its applications in various fields.

thelema418
Messages
131
Reaction score
4
I just finished a course on linear algebra which ended with Jordan Canonical Forms. There were many statements like "Jordan canonical forms are extremely useful," etc. However, we only learned a process to put things into Jordan canonical form, and that was it.

What makes Jordan canonical forms significant? Why would they be useful?
 
Physics news on Phys.org
Indeed, Jordan canonical form is very useful because
1) you cannot always diagonalize a matrix. This acts as a replacement for it.
For example, if you need to solve a system of linear differential equations, and cannot diagonalize the matrix of the system, then you put it into Jordan canonical form, and this leads to a solution (see textbooks).
Similarly, if you have a system of linear recurrence equations, and you cannot diagonalize the system, you put it into Jordan canonical form and solve the system. In brief, each time you would use diagonalization and you cannot diagonalize, use Jordan canonical form.
2) Theoretical aspects : this is used in several arguments in abstract mathematics
 
Do you accept that doing calculations (theoretcial or practical) with matrices is useful? If you accept that idea then one explanation is that Jordan canonical form makes doing calculations easier.

For example, suppose we have a matrix A and we want to compute A^{10}. Suppose you find a matrix B such that B^{-1} A B = J where J is in Jordan cannonical form. To compute J^{10} you only have to raise the Jordan blocks to the 10th power. They individual blocks are usually smaller matrices than the matrix A. From J^{10} you can compute A^{10} from A^{10} = B J^{10} B^{-1}

You can think of Jordan cannonical form as a generalization of the the concept of a "diagonal matrix". It's easy to to multiplications and find inverses (if they exist) for diagonal matrices. If you can find a invertible matrix B that makes matrix B^{-1} A B = D where D is diagonal matrix then you can find A^{10} by the procedure described above. To compute D^{10} you only have to take the 10th power of each diagonal entry.

Why not use diagonal matrices and not worry about Jordan cannonical form? - it's because you can't always find an invertible matrix B such that B^{-1} A B is a diagonal matrix. The Jordan Cannonical form is important because it is the most nearly diagonal format that you can get reliably..

In applications of matrix algebra, expressions of the form B^{-1} (...stuff...) B often have a physical interpretation. The indicate "Change the coordinate system. Do the "stuff" operation in the new coordinate system.. Then change the coordinates of the result back to the original coordinate system. (It's analogous to reading a problem posed in cartesian coordinates, changing to polar coordinates to solve it and then changing back to cartesian coordinates to report the answer.)

If you want to compute the matrix product A^2 C^3 then you can daydream about finding an invertible matrix B such that B^{-1} A B and B^{-1} C B are both diagonal. You use B to change coordinates, you do the computation in the coordinate system where the matrices are diagonal. Then use B^{-1} to change back to the original coordinate system.

You can't always find a matrix B that simultaneously diagonalizes two matrices. You can't always find a matrix B that changes coordinates so two matrices have the same Jordan Block structure. It's important to know special cases when you can. So it's of interest to have theorems that say "If ...so-and-so then the two matrices can be simultaneously diagonalized" or "put in compatible Jordan block form". An even more general concept of expressing matrices in a convenient format is the "singular value decomposition" - but's that's a different topic.
 
Just for clarification, this part of the class did not revolve around a textbook. There was just a handout about how to put a matrix into Jordan canonical form using an algorithm for finding the dot representation of the cycles. So, I don't know how this connects to textbook problems or to real world problems.

I do work with partial and ordinary differential equations that often involves using matrix calculations, but I do not see how this would connect. I understand that if I did not have a calculator, that the process of computing A raised to a power on a similar matrix B could be easier. Yet, calculators and computers can calculate large matrixes raised to a power quickly using the standard left multiplication procedure.

Is this procedure efficient? Or, is it just worthy as a trick to handle situations where you have to calculate by hand?
 
thelema418 said:
I do work with partial and ordinary differential equations that often involves using matrix calculations, but I do not see how this would connect.

Then perhaps you have encountered the exponential function e^{tA} where A is a matrix.
If you put a matrix in diagonal or Jordan cannonical form, you can often see what the result of long calculation will be without doing it - for example, taking the limit of a sequence of matrix operations. The function e^{tA} is defined as a limit of an infinite series involving powers of the matrix A.

There are many numerical algorithms for matrices. If ordinary matrix multiplication met the needs of the world and its computers then many of them would be unnecessary.
 
The problem is not always to compute something, but to obtain an analytic closed expression of something; For example, you may compute the Fibonacci sequence for a given initial condition up to a certain step n, but this does not provide you a general formula for the sequence, and hence you can not obtain approximations, general theorems etc. This is a general philosophical question that is important to understand in mathematics : the fact that you have an algorithm to obtain the solutions of a problem does not mean at all that you have entirely understood it, nor that this provides the tools to deal with questions related to the problem. For example, Lagrange was the first to give the algorithmic solution of the binary quadratic equations in numbers, but this gives nothing about question like "in how many way is it possible to divide a prime numbers in two squares" and an infinity of other questions. This is why Gauss wrote his famous Disquisitiones, where he studied the theoretical aspects of the binary quadratic equations in numbers.
 
To see the connection between differential equations and Jordan forms, recall that a linear constant coefficient differential operator is a product of powers like (D-a)^n, where D stands for differentiation, and a is multiplication by the constant a. Then consider the kernel of such a basic operator acting on the space of smooth functions, e.g. for n = 5, and look at the matrix for D in an essentially standard basis for this space. Note it is already in jordan form:

E.g.: The matrix of D on ker(D-a)^5, with basis {(e^at)(t^r/r!)},0≤r≤4, is:

| a 1 0 0 0 |

| 0 a 1 0 0 |

| 0 0 a 1 0 |

| 0 0 0 a 1 |

| 0 0 0 0 a |, a classic example of a Jordan block, i.e. a “stretch plus a shift”.
 

Similar threads

  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 11 ·
Replies
11
Views
5K
  • · Replies 5 ·
Replies
5
Views
1K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 3 ·
Replies
3
Views
1K
  • · Replies 11 ·
Replies
11
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K