Diagonalization and similar matrices

Click For Summary

Discussion Overview

The discussion focuses on the diagonalization of matrices, specifically addressing the conditions under which a matrix is diagonalizable, the role of eigenvalues and eigenvectors, and the implications of their order in the diagonal matrix. It encompasses theoretical aspects of linear transformations and eigenvalue problems.

Discussion Character

  • Exploratory
  • Technical explanation
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • One participant describes the process of determining if a matrix is diagonalizable by finding eigenvalues and eigenvectors, noting that linear independence of eigenvectors indicates diagonalizability.
  • The same participant questions the significance of the order of eigenvalues when constructing the diagonal matrix, suggesting that different orders could yield different diagonal matrices.
  • Another participant confirms that the order of eigenvalues does not affect the diagonalizability of the matrix, as long as a diagonal matrix is formed.
  • A further contribution clarifies that the order of eigenvalues matters in the construction of the matrix U and its inverse, as U consists of eigenvectors corresponding to the eigenvalues in the same order as they appear in the diagonal matrix D.
  • One participant agrees with this clarification, reinforcing that U's columns must align with the order of eigenvalues in D.

Areas of Agreement / Disagreement

Participants generally agree that the order of eigenvalues does not affect the diagonalizability of the matrix, but they note that it does influence the construction of the matrices U and D. There is no consensus on the implications of choosing different eigenvectors, as this aspect remains less explored.

Contextual Notes

Participants express uncertainty regarding the implications of choosing arbitrary eigenvectors from the eigenspace and the potential effects on the diagonalization process.

Who May Find This Useful

This discussion may be useful for students and practitioners in linear algebra, particularly those interested in matrix diagonalization, eigenvalue problems, and linear transformations.

stripes
Messages
262
Reaction score
0
So when dealing with a linear transformation, after we have computed the matrix of the linear transformation, and we are asked "is this matrix diagonalizable", I begin by finding the eigenvalues and eigenvectors using the characteristic equation.

Once I have found eigenvectors, if I see these vectors are linearly independent, then a theorem tells us the matrix of the linear transformation is, in fact, diagonalizable. Another theorem tells us that this matrix, D, has entries along its main diagonal that are the eigenvalues of the matrix of the linear transformation.

My questions are as follows: what about the order of these eigenvalues? When I solve the characteristic equation, I can order the roots however I want. This means that the matrix of the transformation is diagonalizable to up to n! different diagonal matrices (actually, exactly n! different matrices if we consider complex roots, since all polynomials of degree n contain exactly n roots). My other question is: when choosing eigenvectors that correspond to eigenvalues, can we just choose any vector in the eigenspace, since often times, our eigenvector contains arbitrary constants that can be any real number? I will illustrate my question with an example:

For the matrix A =
\left[ \begin{array}{ccc}<br /> 1 &amp; 1 \\<br /> -2 &amp; 4 \end{array} \right]

Solving for the characteristic polynomial, we have

\lambda = 2, \lambda = 3

whose associated eigenvectors are

\left[ \begin{array}{ccc}<br /> 1 \\<br /> 1 \end{array} \right]

and

\left[ \begin{array}{ccc}<br /> 1 \\<br /> 2 \end{array} \right]. These eigenvectors were chosen with a random scalar constant; can we choose any other eigenvectors?

These two vectors are linearly independent, thus the matrix A can be diagonalized. We see that matrix A is similar to matrix D, where D =

\left[ \begin{array}{ccc}<br /> 2 &amp; 0 \\<br /> 0 &amp; 3 \end{array} \right].

Could we not have stated the roots in a different order, and then eventually conclude that A is similar to D, where D =

\left[ \begin{array}{ccc}<br /> 3 &amp; 0 \\<br /> 0 &amp; 2 \end{array} \right]?

I'm confused as to what order we are to choose, if it even matters (which it seems it does. Sorry if this is a "homework type question", but I think my question is more general and not a specific textbook style question. Sorry if my question is confusing. Any help is appreciated.
 
Physics news on Phys.org
stripes said:
Could we not have stated the roots in a different order, and then eventually conclude that A is similar to D, where D =

\left[ \begin{array}{ccc}<br /> 3 &amp; 0 \\<br /> 0 &amp; 2 \end{array} \right]?

Yeah sure. That's alright to. As long as you end up with a diagonal matrix. There is no preferred order for the eigenvalues.
 
The order in which you put the eingenvalues in the diagonal of D only matter in the construction of U and U inverse where:

A = U D U^-1

Each possible D will have its own U and U inverse pair associated with it.

edit: To be clearer, the way you'd construct U, and D (U inverse is obvious) is: U is the nxn matrix whose columns are the eigenvectors of A in whichever order you wish, and D is the nxn diagonal matrix whose entries are the corresponding eigenvalues to the eigenvectors from U (in the same order).
 
Last edited:
Yes that is right, U's columns are the eigenvectors in the order of the eigenvalues along the main diagonal of D. Thanks everyone for your help.
 

Similar threads

  • · Replies 33 ·
2
Replies
33
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 14 ·
Replies
14
Views
4K
  • · Replies 14 ·
Replies
14
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 3 ·
Replies
3
Views
1K
  • · Replies 2 ·
Replies
2
Views
2K