Characteristic Eq for Matrix problem

Click For Summary
The discussion centers on proving that the coefficient of x in the expansion of det(x*I - A) equals the sum of the determinants of the Aii minors for an n x n matrix A. A user expresses difficulty in deriving this result, attempting various methods including the definition of determinants and Leibniz rule, but finds inconsistencies when Aii is not an eigenvalue. They also explore row permutations and their impact on minors, questioning the connection between diagonal forms and the original matrix. Suggestions are made to consider cases where A is diagonal or diagonalizable, indicating that these forms may simplify the proof. The conversation highlights the complexity of relating the determinant properties of minors back to the original matrix structure.
Zoe-b
Messages
91
Reaction score
0

Homework Statement


Where A is an n x n matrix and I is the n x n identity
In the expansion of det(x*I-A), show that the coefficient of x is equal to the sum from i = 1 to n of the determinant of the Aii minor. (where Aii = the submatrix of A formed by deleting row i and column i)


Homework Equations


if X1, X2, ... Xn are the eigenvalues of A then the coefficient of x is equal to the sum from i = 1 to n of (product of eigenvalues)/Xi.


The Attempt at a Solution


I have tried quite a few different ways of doing this but getting nowhere..
using the definition of a determinant as a sum of products over permuations of Sym (n) and then using the Leibnix rule for integrals I seem to find that (where a11 is the top left entry of A and so on)
d(det(x*I-A)/dx = sum (from i =1 to n) of det(x*I-A)/(x-aii)

The constant coefficient of this will be equal to the x coefficient of det(x*I-A) but the RHS only cancels directly to what I want if aii is an eigenvalue, which obviously isn't always the case. I tried to use row permutations but this obviously changes the Aii minors and so doesn't seem to work. Any ideas? I was trying to follow the proof on

http://www.sciencedirect.com/science?_ob=ArticleURL&_udi=B6TY9-4H5MYJD-1&_user=126524&_coverDate=06%2F30%2F2006&_rdoc=1&_fmt=high&_orig=search&_origin=search&_sort=d&_docanchor=&view=c&_searchStrId=1630570352&_rerunOrigin=google&_acct=C000010360&_version=1&_urlVersion=0&_userid=126524&md5=26cc582e2a0f1358d2f4b36931f4954e&searchtype=a

(sorry for the long link!) but the notation is a bit beyond me. Am I on the right track at all or is there a better way to attempt this? Sorry for not including full working but without the maths language its near impossible to be clear.
Thanks
 
Physics news on Phys.org
Can you prove it for when A is diagonal
 
Yes I understand for when it is diagonal but altho row/column operations of form (row i -> row i + d*row j) won't change the determinant of the matrix, they will change the determinant of the minors.. I don't see how to relate the diagonal form back to the original matrix.
 
How about for a diagonaliable matrix? That is the key here.
 
Hmmmn I'm still pretty clueless but I'll give it another go tomorrow! Thanks
 
Question: A clock's minute hand has length 4 and its hour hand has length 3. What is the distance between the tips at the moment when it is increasing most rapidly?(Putnam Exam Question) Answer: Making assumption that both the hands moves at constant angular velocities, the answer is ## \sqrt{7} .## But don't you think this assumption is somewhat doubtful and wrong?

Similar threads

Replies
3
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
Replies
2
Views
5K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 6 ·
Replies
6
Views
5K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 1 ·
Replies
1
Views
5K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 1 ·
Replies
1
Views
4K