How do I prove this? [Linear Algebra]

In summary: I can't summarize the conversation because it contains a lot of technical details and proofs. In summary, the conversation discusses the equivalence between three statements: the existence of a linearly independent set of vectors, the characteristic polynomial being equal to the minimal polynomial, and the existence of a specific basis for a linear operator. The participants also discuss the difficulty of proving the implications in both directions and the usefulness of certain facts about the characteristic and minimal polynomials. They ultimately conclude that the statements are in fact equivalent and provide a proof by showing that the characteristics polynomials of [T]_{\beta} and T are the same.
  • #1
AKG
Science Advisor
Homework Helper
2,567
4
Let [itex]T : V \to V[/itex] be a linear operator on an n-dimensional vector space V. Assume that the characteristic polynomial of T splits over F (the field underlying V). Prove that the following are equivalent:
  1. There exists a vector [itex]x \in V[/itex] such that [itex]\{x,T(x),T^2(x),\dots ,T^{n - 1}(x)\}[/itex] is linearly independent.
  2. The characteristic polynomial of T is equal to [itex](-1)^n[/itex] times the minimal polynomial of T.
  3. There exists a basis [itex]\beta[/itex] such that

    [tex][T]_{\beta} = \left (\begin{array}{ccccc}0 & 0 & \dots & 0 & (-1)^{n - 1}a_0\\1 & 0 & \dots & 0 & (-1)^{n - 1}a_1\\0 & 1 & \dots & 0 & (-1)^{n - 1}a_2\\ \vdots & \vdots & \ddots & \vdots & \vdots \\0 & 0 & \dots & 1 & (-1)^{n - 1}a_{n - 1}\end{array}\right )[/tex]

    where [itex](-1)^nt^n + a_{n - 1}t^{n - 1} + \dots + a_1t + a_0[/itex] is the characteristic polynomial of T.
    (Hint: For proving that (2) implies (3), it is helpful to show that the Jordan canonical form of T is the same as the Jordan canonical form of the matrix given in part (3)).
I can prove (1) implies (2). In my book, the section on Jordan canonical form comes before the section on minimal polynomials, and then the section on minimal polynomials says very little about Jordan canonical form, so the hint given to prove (2) implies (3) isn't helping me too much. Any hints and suggestions would be helpful, thanks.
 
Last edited:
Physics news on Phys.org
  • #2
Can this even be proved? In my book, a proof for (1) implies (2) is basically given (which is why I can do it), but it says nothing along the lines of (2) implies (1). That is, if it is in fact true that (1) iff (2), why would my book only prove (1) implies (2) and not even suggest that the converse is true?

Furthermore, I can't seem to figure out what the minimal polynomial tells you about the Jordan canonical form at all, so I can't see why (2) would imply (3). The only possible useful facts I know are that the Jordan canonical form of an operator is the same as the Jordan canonical form of the matrix of that operator with respect to some basis, and that if two operators', S's and T's, characteristic polynomials split, then they have the same Jordan canonical form if and only if they are similar to each other, i.e. there exists an invertible operator U such that [itex]S = U^{-1}TU[/itex]. Let's call the matrix labeled [itex][T]_{\beta}[/itex] A, i.e. [itex]A = [T]_{\beta}[/itex]. Let [itex]\beta \prime[/itex] be a Jordan canonical basis for T. Then [itex][T]_{\beta \prime}[/itex] is the Jordan canonical form for T. Now, if I compute the characteristic polynomial of A, I suspect I will get it to look something like the characteristic polynomial of T, and therefore I know it will split. Now, if I can show that the Jordan canonical form of A is the same as [itex][T]_{\beta \prime}[/itex], then that's implies that A is similar to [itex][T]_{\beta \prime}[/itex]. If this is true, then [itex]A = U^{-1}[T]_{\beta \prime}U[/itex]. U can be seen as a change of basis matrix, specifically, a change of basis matrix from some basis [itex]\beta[/itex] to [itex]\beta \prime[/itex]. This implies that there is some basis [itex]\beta[/itex] such that [itex][T]_{\beta} = A[/itex], as required.

The problem with the above is getting the Jordan canonical form of the operator T satisfying condition (2). Like I said, all I have, as far as I can tell, are some facts about the characteristic polynomial and minimal polynomial of T, but I don't think I have enough to say anything about it's Jordan canonical form. Is there something I'm missing.

And I haven't even looked at how I would go about proving (3) implies (1). Glancing at it, I'm even more lost for that part than I am for the part proving (2) implies (3).

Please! Any help will be appreciated!
 
  • #3
I remember freshman year fun stuff!
You are really good with all the fancy symbol markups. I am not, but I hope I can still be clear.

It seems you define the characteristic polynomial as det(A-I*t) so we have all the (-1)^n stuff and that you define the minimal polynomial to be monic. I hope when you talk about what section are before others that you read all of them, I assume minimal polynomial is defined as
a monic poly mp so that mp(T)=0 and if p(T)=0 for some poly p, then mp divides p. I we use characteristic of T splits over F only to assure T has n eigen values in F (counting multiplicity).

I think
3=>1=>2=>3 should do it

3=>1
let x=e(1) the vector [1,0,...,0]
then [tex][T]_{\beta}[/tex]^k*e(1)=e(k+1) k=0,...,n-1
(this is clear from the simple basis)
clearly {e(1),...,e(n)} is linearly independent
so 1 is true in the basis B hence true in any basis
qed

1=>2
1 implies that no nontrivial polynomials in T of order less than n are equal to the 0 operator. So deg(minimal polynomial)=n
deg(characteristic polynomial)=n so they are equal up to multiplication by a constant. By the definitions the constant is (-1)^n
qed

2=>3
we will show [tex][T]_{\beta}[/tex] and T have the same jordan conical form, thus they are similar, hence a basis between them exist. We know [tex][T]_{\beta}[/tex] will have all of its Jordan blocks full. That is if lamda is an eigenvalue of multiplicity m Nullity(([tex][T]_{\beta}[/tex]-lamda*I)^k)=k k=0,1,...,m, and the jordan block is J_m(lamda). We know this because even though we do not yet know that the basis B will exist given (2) we do know from above that (1) and (2) hold (for [tex][T]_{\beta}[/tex]) if [tex][T]_{\beta}[/tex] is defined as it is in (3) Thus [tex][T]_{\beta}[/tex] and T will have the same Jordan form iff they have the same characteristic function.
we see
[tex][T]_{\beta}[/tex]^n*e(1)=(-1)^k Sum[a(k)e(k+1),{k,0,n-1}]
putting
e(k+1)=[tex][T]_{\beta}[/tex]^k+*e(1) k=0,...,n-1 as before
we see
((-1)^n [tex][T]_{\beta}[/tex]^n+Sum[a(k)[tex][T]_{\beta}[/tex]^k,{k,0,n-1}])*e(1)=0
appying [tex][T]_{\beta}[/tex]^k k=0,...,n-1 to both sides and knowing polynomials in [tex][T]_{\beta}[/tex] commute we have
((-1)^n [tex][T]_{\beta}[/tex]^n+Sum[a(k)[tex][T]_{\beta}[/tex]^k,{k,0,n-1}])*[tex][T]_{\beta}[/tex]^k*e(1) k=0,...,n-1
or
[tex][T]_{\beta}[/tex]^n+Sum[a(k)[tex][T]_{\beta}[/tex]^k,{k,0,n-1}])*e(k)=0 k=1,...,n
{e(1),...,e(n)} is a basis so
[tex][T]_{\beta}[/tex]^n+Sum[a(k)[tex][T]_{\beta}[/tex]^k,{k,0,n-1}])=0
thus the two characteristic polys agree [tex][T]_{\beta}[/tex] and T are similar (since same Jordan form) so the basis exists
 
Last edited:
  • #4
Thanks anyways, I did it already. I wasn't able to do it before because I needed to know something that hadn't been taught at that point, nor was it in the book, namely "We know [itex][T]_{\beta}[/itex] will have all of its Jordan blocks full."
 

1. How do I prove that two matrices are equal?

To prove that two matrices are equal, you need to show that all corresponding elements in each matrix are equal. This can be done by comparing the values in each position and showing that they are the same. You can also use algebraic properties such as the commutative and associative properties to rearrange the elements and show that they are equal.

2. How do I prove that a matrix is invertible?

A matrix is invertible if its determinant is non-zero. To prove this, you can calculate the determinant of the matrix and show that it is not equal to zero. Additionally, you can show that the matrix has a unique solution when solving for the inverse, which further proves its invertibility.

3. How do I prove that a linear transformation is one-to-one?

A linear transformation is one-to-one if it maps distinct elements to distinct elements. To prove this, you can show that the kernel (the set of vectors that map to the zero vector) of the transformation is only the zero vector. This means that every vector maps to a unique output, proving that the transformation is one-to-one.

4. How do I prove that a set of vectors is linearly independent?

A set of vectors is linearly independent if none of the vectors can be written as a linear combination of the others. To prove this, you can set up a system of equations using the vectors as the variables and solve for the coefficients. If the only solution is when all coefficients are equal to zero, then the set of vectors is linearly independent.

5. How do I prove that a set of vectors spans a vector space?

A set of vectors spans a vector space if any vector in the space can be written as a linear combination of the given vectors. To prove this, you can set up a system of equations using the vectors as the variables and solve for the coefficients. If there exists a solution for any vector in the space, then the set of vectors spans the vector space.

Similar threads

  • Introductory Physics Homework Help
Replies
3
Views
734
  • Calculus and Beyond Homework Help
Replies
24
Views
784
  • Calculus and Beyond Homework Help
Replies
4
Views
955
  • Introductory Physics Homework Help
Replies
2
Views
984
Replies
34
Views
2K
  • Calculus and Beyond Homework Help
Replies
7
Views
402
  • Linear and Abstract Algebra
Replies
3
Views
1K
  • Linear and Abstract Algebra
Replies
3
Views
1K
Replies
27
Views
1K
  • Programming and Computer Science
Replies
9
Views
1K
Back
Top