On the definition linear independence and dependence

cartonn30gel
Messages
68
Reaction score
0
This is not really a homework problem but it relates to a number of problems, so I thought this would be the most appropriate place to post it.

Homework Statement



The basic question is about how we define linear dependence in a vector space. For a vector space over some field \mathbb{F}, we know that the vectors v_1,v_2,...,v_n are linearly independent if the only solution to a_1v_1+a_2v_2+...+a_nv_n= 0 is a_1=a_2=...=a_n=0. And if some list of vectors are not linearly independent, they are linearly dependent. This means if we can find constants b_1,..., b_n that are NOT ALL ZERO where b_1v_1+b_2v_2+...+b_nv_n=0, these vectors are linearly dependent.

Now when we think about the polynomial space (let's say over some field) P_m(\mathbb{F}), we need to reconsider these definitions right? I have never seen a different defintion of linear independence for the polynomial space but I'm assuming it would be like this:
The vectors p_1(z),...,p_m(z) are linearly independent if the only solution to c_1p_1(z)+...+c_mp_m(z)=0 FOR ALL z is c_1=...=c_m=0.

And linear dependence would be this: If we can find constants c_1, ..., c_m that are not all zero FOR SOME z where c_1p_1(z)+...+c_mp_m(z)=0, these vectors are linearly dependent.

Notice that if linear dependence on the polynomial space is defined this way, it is actually the exact negation of linear independence in the polynomial space. But if linear dependence is defined in the similar way but FOR ALL z, then it is NOT the exact negation of linear independence in the polynomial space. And then, we start encountering vectors that are neither linearly independent nor linearly dependent.

Please try to shed some light on this. Is my given definitions for linear independence and dependence for the polynomial space correct?

Homework Equations


Known definitions of linear independence and dependence are given above.

The Attempt at a Solution


There isn't really a solution. It is just a question about definitions.
 
Last edited:
Physics news on Phys.org
When you see an equation like c_{1}f_{1}(z) + ... + c_{n}f_{n}(z) = c where c is some constant, then r is taken to be the function r(z) = r. This takes care of all of the details because f(z) = 0(z) iff f is zero on all z so if it is non-zero on at least one z then it's not the zero function.
 
Last edited:
Looks fine to me. You seem to get the fine point that for a linearly independent set of vectors, this equation has only one solution.
a_1v_1+a_2v_2+...+a_nv_n= 0

The same equation for an arbitrary set of vectors, whether linearly independent or linearly dependent, always has the solution a1 = a2 = ... . an = 0. The key difference, and one that students have a hard time with, is whether this solution is the only one.

Your definition for linear independence in a function space is fine, too. A key point there is that the equation has to hold for all values of the variable.
 
aPhilosopher,
I can't see how what you said is relevant in this case. I am talking about the specific coefficients that multiply the polynomials. Are they functions of z as well? (I think they aren't because it is called "scalar" multiplication) Can you give some more details please?

And Mark44,
so linear dependence in the vector space of polynomials is still the exact negation of linear independence. And its definition of it is:

If we can find constants c_1, ..., c_m that are not all zero FOR SOME z where c_1p_1(z)+...+c_mp_m(z)=0, these vectors are linearly dependent.

Correct?

If it is correct, I will post an example question and my solution to it making use of this fact.
 
Last edited:
cartonn30gel said:
aPhilosopher,
I can't see how what you said is relevant in this case. I am talking about the specific coefficients that multiply the polynomials. Are they functions of z as well? (I think they aren't because it is called "scalar" multiplication) Can you give some more details please?

And Mark44,
so linear dependence in the vector space of polynomials is still the exact negation of linear independence. And its definition of it is:

If we can find constants c_1, ..., c_m that are not all zero FOR SOME z where c_1p_1(z)+...+c_mp_m(z)=0, these vectors are linearly dependent.

Correct?
"... these functions are linearly dependent." Minor point, as functions in a function space are the counterparts of vectors in a vector space.
cartonn30gel said:
If it is correct, I will post an example question and my solution to it making use of this fact.
 
cartonn30gel said:
aPhilosopher,
I can't see how what you said is relevant in this case. I am talking about the specific coefficients that multiply the polynomials. Are they functions of z as well? (I think they aren't because it is called "scalar" multiplication) Can you give some more details please?

I noticed that my notation was a bit confusing so I edited my previous post. the coefficients c_{i} are constant but the polynomials f_{i} are not. So when you take the linear combination \sum_{i} c_{i}f_{i}, the result is a function. If you set this equal to something, that something had better be a function as well. So functions f_{i} are linearly dependent if there exists c_{i}, not all zero, such that \sum_{i} c_{i}f_{i} is the zero function. A function is the zero function if and only if it is zero on every element of it's domain. It is not the zero function if it is non-zero on at least one element of it's domain.

The point is that all the business about "at least"/"for all" is already taken care of in the definition of the zero function and the realization that zero means the zero function in this context.

That being said, it is a good observation.
 
There are two things I don't understand about this problem. First, when finding the nth root of a number, there should in theory be n solutions. However, the formula produces n+1 roots. Here is how. The first root is simply ##\left(r\right)^{\left(\frac{1}{n}\right)}##. Then you multiply this first root by n additional expressions given by the formula, as you go through k=0,1,...n-1. So you end up with n+1 roots, which cannot be correct. Let me illustrate what I mean. For this...

Similar threads

Back
Top