Problem on linear independence and matrices

vince89
Messages
6
Reaction score
0
Can I ask for some help?

Suppose that {v1,v2...vn} is a linearly independent set of vectors and A is a singular matrix.
Prove or disprove: The set {Av1, Av2, ...Avn} is linearly independent.
 
Physics news on Phys.org
How can you relate two different basis?
 
the vectors are the same. and they're not bases, they're just sets of vectors
 
But the vetors of the first set are linearly independent, so they form a basis. If the second set has linearly independent members it should also be a base. Correct?
 
but doesn't the singular matrix change the property of linear independence?
 
So the members of the second set can not be independent! :smile:
 
hmm...i don't quite get it. can you write a rough proof. :) thanks a lot!
 
Ok!
The first set forms a basis since the vectors are linearly independent.

Suppose now that the members of the 2nd set are linearly independent, thus they also form a basis. Two basis are related by a non-singular matrix, but in our case they related by a singular one. Thus the members of the 2nd are not linearly independent.

What about that?
 
how does the singular matrix change the linear independence of the basis?
 
  • #10
Let me show it with equations.

Call the vectors of the 2nd set \bar{v}^\alpha, then

\bar{v}^\alpha=A^\alpha_\beta\,v^\beta

In order for \bar{v}^\alpha to be linearly independent, it must hold that

\lambda_\alpha \, \bar{v}^\alpha=0\Rightarrow \lambda_{\alpha}=0 \quad \forall \alpha

Now we have

\lambda_\alpha \, \bar{v}^\alpha=0\Rightarrow \lambda_\alpha \, A^\alpha_\beta\,v^\beta=0 \Rightarrow \lambda_\alpha \, A^\alpha_\beta=0

The last equality hols because v^\alpha[/tex] are linearly independent.<br /> <br /> This is a n\times n <i>homogeneous</i> system for the unknows \lambda_\alpha. In order for that to have only the trivial solution it must hold det(A)\neq 0. But det(A)= 0 so there is a solution for \lambda_\alpha besides the trivial one. Thus the vectors \bar{v}^\alpha are dependent.
 
Last edited:
  • #11
if there is a trivial solution, what does it imply? :) thanks btw.
 
  • #12
How about just trying for a counter example? What is the simplest singular linear transformation you know?
 
  • #13
Trivial solution means \lambda^\alpha=0, which would cause \bar{v}^\alpha to be independent
 
  • #14
Hello,

I hope I am not spoiling the fun, but I think things are getting confused.

Anything could happen here. Since we are not requiring that {v_1, ..., v_n} is a basis, here are two examples. Take any set of linearly independent vectors and let A = 0. Then of course their images are not linearly independent. On the other hand, take A to be some nonzero matrix, and let v_1 be any vector not in its kernel. Then {v_1} is linearly independent, and so is {Av_1}.
 
  • #15
masnevets said:
Hello,

I hope I am not spoiling the fun, but I think things are getting confused.

Anything could happen here. Since we are not requiring that {v_1, ..., v_n} is a basis, here are two examples. Take any set of linearly independent vectors and let A = 0. Then of course their images are not linearly independent. On the other hand, take A to be some nonzero matrix, and let v_1 be any vector not in its kernel. Then {v_1} is linearly independent, and so is {Av_1}.

The OP was:

Suppose that {v1,v2...vn} is a linearly independent set of vectors and A is a singular matrix.

Since the v_i are linearly independent, then they form a basis. Assuming of course, the dimension of the vector space is n. :smile:
 
  • #16
All you really need is a 'counterexample'. If {v1, v2, ..., vn} is a set of independent vectors, and A is the linear transformation that takes every v into 0, what can you say about {Av1, Av2, ..., Avn}?
 
  • #17
whats your definition of singular? if it means the columns are not independent, you are done. at least assuming you know the basic theory of dimension.
 
  • #18
Rainbow Child said:
The OP was:



Since the v_i are linearly independent, then they form a basis. Assuming of course, the dimension of the vector space is n. :smile:
Yes, and masnevets' point was that there is no reason to assume that! In any case, his point was an extension of what I said: Suppose {v1, v2, ...} is a set of independent vectors and A is the zero operator (Av= 0 for all v).
 
Back
Top