Non-zero determinant iff matrix is invertible.

ksm100
Messages
5
Reaction score
0

Homework Statement


Given that A is any 2x2 matrix show that it is invertible if and only if det(A) \neq 0.

Homework Equations





The Attempt at a Solution


If A is invertible then we know there exists an inverse matrix, say B, such that AB = BA = I.
It follows that det(AB)=det(BA)=det(I), and we know that det(I) = (1*1) - (0*0) = 1, so
det(AB) = det(A)det(B) = 1 implies both det(A) and det(B) are both nonzero.

However, I'm unsure how to show the converse.

If we suppose det(A) is not equal to 0, we know that no rows/columns of A are all zero, no two rows/columns are equal, and one row/column is not a multiple of the other.

I'm stuck here.. if anyone could help I'd really appreciate it!
 
Physics news on Phys.org
How would you write A-1 in terms of det(A) and adj(A)?
 
Do you have any explicit formulas* for the inverse?

If not, then if you wrote down a generic matrix (its entries are variables), do you have an algorithm to compute* the inverse?

If not, can you write down another generic matrix at least solve* the system of equations that says "this matrix is the inverse of the other one"?

*: Paying careful attention to when it does and doesn't work? e.g. if you would divide by the expression (b-a), you should keep track of the fact you're assuming b-a is nonzero. (And then consider the case b=a separately)
 
A^(-1) = [1/det(A)]*adj(A)

So I can just say that because I know that det(A) isn't zero, 1/det(A) is defined and therefore A^(-1) exists?
 
ksm100 said:
A^(-1) = [1/det(A)]*adj(A)

So I can just say that because I know that det(A) isn't zero, 1/det(A) is defined and therefore A^(-1) exists?

Well I think that would work. I am not sure what sort of proof you are expected to give though.
 
Use the fact that an n*n matrix is invertible if and only if its rank is n.
 
There are two things I don't understand about this problem. First, when finding the nth root of a number, there should in theory be n solutions. However, the formula produces n+1 roots. Here is how. The first root is simply ##\left(r\right)^{\left(\frac{1}{n}\right)}##. Then you multiply this first root by n additional expressions given by the formula, as you go through k=0,1,...n-1. So you end up with n+1 roots, which cannot be correct. Let me illustrate what I mean. For this...
Back
Top