Proving Similarity of Non-Diagonalizable Matrices

  • Thread starter Thread starter Lily@pie
  • Start date Start date
  • Tags Tags
    Matrix
Click For Summary
To prove the similarity of a non-diagonalizable 2x2 matrix A to the form [[Ω,1],[0,Ω]], it is established that A and the matrix L share the same eigenvalue, Ω. The discussion emphasizes that since A is not diagonalizable, it cannot have two distinct eigenvalues, leading to the conclusion that both matrices must have the same eigenvalue structure. The participants explore the implications of generalized eigenvectors and the Jordan normal form, noting that while direct use of the Jordan form is not permitted, understanding its principles is essential for proving similarity. The conversation also touches on the need to demonstrate the existence of a linearly independent vector w that, along with the eigenvector v, forms a basis for R², ultimately confirming the similarity condition. The thread concludes that while the proof is complex, it hinges on establishing the relationship between the eigenvalues and the ranks of the corresponding eigenspaces.
Lily@pie
Messages
108
Reaction score
0

Homework Statement


A is a 2 by 2 real matrix which cannot be diagonalized by matrix P. Prove that there is a invertible P such that P-1AP=[[Ω,1][0,Ω]]

2. The attempt at a solution
I didn't know how to do this so I tried the following.

Since we need to prove there is a invertible P such that P-1AP=[[Ω,1][0,Ω]], this means that we need to prove A and L:=[[Ω,1][0,Ω]] are similar matrix.

So we need to show A and L have the same eigenvalue whenever A is not diagonalizable.

det (xI-L) = (x-Ω)2
eigenvalue of L =Ω

det (xI-A) = x2-(a+d)x+ad-bc
eigenvalue of A = 1/2(a+d+SQRT[(a+d)^2-4(ad-bc)]) or 1/2(a+d+SQRT[(a+d)^2+4(ad-bc)])
Since A is not diagonalizable, it will not have 2 distinct eigenvalues. This implies that (a+d)^2-4(ad-bc) = 0. Hence eigenvalue of A =(x- (a+d)/2)2

Since Ω can be any number, the eigenvalues of A and L will be the same when A is not diagonalizable. This implies that A and L are similar matices. (This is the main part that I am not sure about, it seems wrong)

Therefore, there exist an invertible P such that P-1AP=[[Ω,1][0,Ω]]
 
Physics news on Phys.org
But because I wasn't taught on the Jordan Normal form and my lecturer state that it can be done without it. Will my prove be valid?
 
I think your instructor wants you go through the construction of the Jordan normal form matrix.

If you already knew about Jordan normal form, the problem is trivial. You would recognize L is that form, so there's exists a matrix P, blah blah blah.

On the other hand, if you go through the logic of where these generalized eigenvectors come from, you can show they're independent and therefore they're a basis for R2. Then you can show, for example, what (A-IΩ) has to look like in that basis, and so on. You're not so much using Jordan normal form; you're deriving it.

This way you'll understand why the matrix has the form it does, which is probably what your instructor wants you to learn, rather than just knowing how to write down the matrix without understanding where it came from.EDIT: Of course, I could be wrong. Use your own judgment. :)
 
Last edited:
I will look up Jordan normal form. Just want to ask one question, does having the same eigenvalue implies that the matrices are similar? Because I know that similar matrices implies that they have the same eigenvalue.
 
No, the matrices
\begin{pmatrix} 0 & 0 \\ 0 & 0 \end{pmatrix}and
\begin{pmatrix} 0 & 1 \\ 0 & 0 \end{pmatrix}have the same eigenvalue, but obviously, they can't be similar.
 
Oh... ok...
I still don't really understand the jordan normal form... urgh!

btw, is there any characteristic that we can use to prove that 2 matrix are similar?
 
The only way I know of is to find the matrix P.
 
Aren't 2 matrices similar if they have the same Jordan normal form (using some ordering)?
That is, they are similar if they have the same eigenvalues combined with the same ranks of the corresponding eigen spaces?


Btw, is Ω supposed to be a real number?
Because if so, I don't think it's always possible.
What if there are no real eigenvalues?
Like in:
A=\begin{pmatrix}0 & -1 \\ 1 & 0 \end{pmatrix}
 
  • #10
Ω is not specify to be anything... I presume it can be complex or real...

So will I be able to say since the both of them have the same eigenvalues and ranks to the corresponding eigenspace, so they are similar?
 
  • #11
Lily@pie said:
Ω is not specify to be anything... I presume it can be complex or real...

So will I be able to say since the both of them have the same eigenvalues and ranks to the corresponding eigenspace, so they are similar?

I believe you can indeed say that, but... you would still need to proof it...
I dare to say it because of the theorem of the Jordan normal form, but if you're not supposed to use that...
 
  • #12
oh my... how am I suppose to prove this. I've read so many on jordan normal form. All I know is a non-diagonalized matrix can be written in jordan normal form. and...

Any hints??
 
  • #13
I'm a bit fuzzy on what you can and cannot use.
Which theorems do you have available?

For starters, let me review your step:
"Since A is not diagonalizable, it will not have 2 distinct eigenvalues."

It is true, but how do you know this?
Which theorem are you using here?
Otherwise you should still proof it.
 
  • #14
Because if A is not diagonal, this means that it will have 1 or less eigenvector. Which also implies that it would be impossible to have 2 different eigenvalues as 2 different eigenvalues will lead to 2 eigenvectors...

That's what I used to deduce that...
 
  • #15
Lily@pie said:
Because if A is not diagonal, this means that it will have 1 or less eigenvector. Which also implies that it would be impossible to have 2 different eigenvalues as 2 different eigenvalues will lead to 2 eigenvectors...

That's what I used to deduce that...

Apparently you are using a theorem on the diagonizability of a matrix...

Let me rephrase.

If a 2x2 matrix A has 2 distinct eigenvalues a and b, it also has 2 independent corresponding eigenvectors v and w.
(Why are they independent?)

So Av = av and Aw = bw.
This means that A (\boldsymbol v ~ \boldsymbol w) = (a\boldsymbol v ~ b\boldsymbol w) = (\boldsymbol v ~ \boldsymbol w) \begin{pmatrix}a & 0 \\ 0 & b \end{pmatrix}.

There! :smile:
With P=(v w) we have the (sub)proof that 2 distinct eigen values imply diagonizability.
(Why is P invertible? And why does this imply similarity?)TBH, I haven't worked your problem out (completely) myself yet.
But I think your proof should be something similar as what we did just now.
 
Last edited:
  • #16
I like Serena said:
There! :smile:
With P=(v w) we have the (sub)proof that 2 distinct eigen values imply diagonizability.
(Why is P invertible? And why does this imply similarity?)


So we have therefore proven that not diagonalizable matrix implies that the is no 2 distinct eigenvalues.

P will be invertible because v and w are linearly independent. So this implies similarity between A and [[a,0][0,b]] since P-1AP=[[a,0][0,b]]

What if now we change [[a,0][0,b]] to [[Ω,1][0,Ω]]?? Does this tell us anything...
So A[v w]=[v w][[Ω,1][0,Ω]]
Av=Ωv which implies that v is an eigenvector for A with eigenvalue Ω.
Aw=v+Ωw which... erm, I don't really know
 
  • #17
Lily@pie said:
So we have therefore proven that not diagonalizable matrix implies that the is no 2 distinct eigenvalues.

P will be invertible because v and w are linearly independent. So this implies similarity between A and [[a,0][0,b]] since P-1AP=[[a,0][0,b]]

What if now we change [[a,0][0,b]] to [[Ω,1][0,Ω]]?? Does this tell us anything...
So A[v w]=[v w][[Ω,1][0,Ω]]
Av=Ωv which implies that v is an eigenvector for A with eigenvalue Ω.
Aw=v+Ωw which... erm, I don't really know

Yep.
So if we can prove that we can always find a w, linearly independent of v, such that Aw=v+Ωw, we're basically done.

Btw, we can already say that for any vector w independent of v, that (A-ΩI)w≠0.
(Why?)
 
  • #18
I like Serena said:
Yep.
So if we can prove that we can always find a w, linearly independent of v, such that Aw=v+Ωw, we're basically done.

Btw, we can already say that for any vector w independent of v, that (A-ΩI)w≠0.
(Why?)


because v is an eigenvector which is by definition, a non-zero vector. So (A-ΩI)w≠0 since (A-ΩI)w=v

but how do we show v and w when (A-ΩI)w=v are always linearly independent?
 
  • #19
I just went over the wiki article on Jordan normal forms again:
http://en.wikipedia.org/wiki/Jordan_normal_form

It gives a method to find w (a generalized eigenvector):

From:
Aw=v+Ωw​
we get:
(A-ΩI)w=v
(A-ΩI)2w=(A-ΩI)v=0​

So w is a vector in the kernel of (A-ΩI)2.

The article also gives a proof why this always works.
Perhaps it can be simplified for a 2x2 matrix.
 
  • #20
I like Serena said:
I just went over the wiki article on Jordan normal forms again:
http://en.wikipedia.org/wiki/Jordan_normal_form

It gives a method to find w (a generalized eigenvector):

From:
Aw=v+Ωw​
we get:
(A-ΩI)w=v
(A-ΩI)2w=(A-ΩI)v=0​

So w is a vector in the kernel of (A-ΩI)2.

The article also gives a proof why this always works.
Perhaps it can be simplified for a 2x2 matrix.

I could understand until the fact that v is in the intersection of Range (A-ΩI) & Ker(A-ΩI) and w is a vector in the kernel of (A-ΩI)2.

how to relate to w?? hmm... since w is not in the Ker(A-ΩI), v≠w...
 
  • #21
Lily@pie said:
I could understand until the fact that v is in the intersection of Range (A-ΩI) & Ker(A-ΩI) and w is a vector in the kernel of (A-ΩI)2.

how to relate to w?? hmm... since w is not in the Ker(A-ΩI), v≠w...

Yep.
We need to proof that there exists a w in the kernel of (A-ΩI)2 that is not in the kernel of (A-ΩI).

Note that v is in the kernel of (A-ΩI), so it is also in the kernel of (A-ΩI)2.

We already know that the rank of (A-ΩI) is one.
If we can proof that the rank of (A-ΩI)2 is two, then that implies that w exists, which is sufficient (we don't have to actually find it).


Perhaps an alternate avenue is to start with a generic matrix A and a generic eigenvector v and actually think up a vector w that does the job.
 
  • #22
I like Serena said:
Yep.
We need to proof that there exists a w in the kernel of (A-ΩI)2 that is not in the kernel of (A-ΩI).

Note that v is in the kernel of (A-ΩI), so it is also in the kernel of (A-ΩI)2.

We already know that the rank of (A-ΩI) is one.
If we can proof that the rank of (A-ΩI)2 is two, then that implies that w exists, which is sufficient (we don't have to actually find it).


Perhaps an alternate avenue is to start with a generic matrix A and a generic eigenvector v and actually think up a vector w that does the job.

Can we say that since (A-ΩI) is a 2 by 2 matrix and we found 2 vectors v and w where v≠w, so rank of (A-ΩI)2 is two?

but how does this shows that v and w are independent vectors?
 
  • #23
Lily@pie said:
Can we say that since (A-ΩI) is a 2 by 2 matrix and we found 2 vectors v and w where v≠w, so rank of (A-ΩI)2 is two?

but how does this shows that v and w are independent vectors?

We haven't found w yet.
We need to proof that w exists (and is linearly independent of v).
 
  • #24
You can prove that w isn't a multiple of v, so it follows w is independent of v and they form a basis for R2. Using the fact that (A-ΩI)v=0 and (A-ΩI)w=v, you can write down what matrix represents (A-ΩI) in that basis, i.e. what P-1(A-ΩI)P is where P=(v w).
 
  • #25
vela said:
You can prove that w isn't a multiple of v, so it follows w is independent of v and they form a basis for R2. Using the fact that (A-ΩI)v=0 and (A-ΩI)w=v, you can write down what matrix represents (A-ΩI) in that basis, i.e. what P-1(A-ΩI)P is where P=(v w).

How do you know that you can find a w that satisfies (A-ΩI)w=v?
I'm still stuck at that. :frown:
 
  • #26
(A-ΩI)w has to map to something, and you know that (A-ΩI)2w=0 for all w in R2. (A-ΩI)[(A-ΩI)w] = 0 means (A-ΩI)w is an eigenvector or 0, so ...
 
  • #27
vela said:
(A-ΩI)[(A-ΩI)w] = 0 means (A-ΩI)w is an eigenvector or 0, so ...

Aaaaaah! :smile:
 
  • #28
So (A-ΩI)w is an eigenvector or 0.

If (A-ΩI)w=0, w will be a eigenvector for A? I'm still lost...:cry:
 
  • #29
Lily@pie said:
So (A-ΩI)w is an eigenvector or 0.

If (A-ΩI)w=0, w will be a eigenvector for A? I'm still lost...:cry:

We skipped a step before.

We had established that since A is not diagonalizable, its eigenvalues could not be distinct.
So the eigenvalues are the same (Ω).

However, if there are 2 linearly independent eigenvectors for Ω, A is still diagonalizable (with P as the matrix defined by those 2 eigenvectors).
So if v is an eigenvector, any linearly independent w can not be an eigenvector.
So (A-ΩI)w≠0.
So how must (A-ΩI)w be related to v now?
 
Last edited:
  • #30
I like Serena said:
We skipped a step before.

We had established that since A is not diagonalizable, its eigenvalues could not be distinct.
So the eigenvalues are the same (Ω).

However, if there are 2 linearly independent eigenvectors for Ω, A is still diagonalizable (with P as the matrix defined by those 2 eigenvectors).
So if v is an eigenvector, any linearly independent w can not be an eigenvector.
So (A-ΩI)w≠0.

But having (A-ΩI)w=0 means w is an eigenvector for A. But we know it cannot be 2 distinct eigenvectors, so w can only be a multiple of v. Which shows that v and w are linearly dependent.

But we would hope to show v and w are linearly independent.

(A-ΩI)w≠0 mean (A-ΩI)w is an eigenvector for A?? So w and v are linearly independent??
don't understand ><

Besides that writting P-1(A-ΩI)P where P=[v w] means writing the eigenvalues in the diagonal entry??
 

Similar threads

Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K
Replies
5
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
Replies
4
Views
2K
Replies
9
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K