## Linear independence

Any one plz tell me about the term linear independence?and when we say that the function is linear independent
 PhysOrg.com science news on PhysOrg.com >> City-life changes blackbird personalities, study shows>> Origins of 'The Hoff' crab revealed (w/ Video)>> Older males make better fathers: Mature male beetles work harder, care less about female infidelity
 Recognitions: Science Advisor You need to supply context. Usually linear independence refers to a set of vectors or a set of functions. These things are called linearly dependent if some linear combination = 0. If not then they are linearly independent.
 [A][/A]=[c][/1]x+[c][/2]y[j][/j]+[c][/3]z[k][/k] when we say it is linearly independent and also my friend argue with me that orthonormality implies linear independence but i was not satisfied plz help

Recognitions:

## Linear independence

 Quote by sncum [A][/A]=[c][/1]x+[c][/2]y[j][/j]+[c][/3]z[k][/k] when we say it is linearly independent and also my friend argue with me that orthonormality implies linear independence but i was not satisfied plz help
I don't understand your first line.

However your friend is correct, if two vectors are orthogonal (unless one of the is 0) they are linearly independent. Note that the converse is not true.

Mentor
 Quote by sncum my friend argue with me that orthonormality implies linear independence but i was not satisfied plz help
A set E is said to be linearly independent if for all finite subsets ##\{x_1,\dots,x_n\}\subset E## and all ##a_1,\dots,a_n\in\mathbb C##,
$$\sum_{k=1}^n a_k x_k=0\quad \Rightarrow\quad a_1=\dots=a_n=0.$$
Suppose that E is orthonormal. Let ##\{e_k\}_{k=1}^n## be an arbitrary finite subset of E, and suppose that ##\sum_{k=1}^n a_k e_k=0##. Then for all ##i\in\{1,\dots,n\}##,
$$0=\langle e_i,0\rangle=\langle e_i,\sum_k a_k e_k\rangle=\sum_k a_k\langle e_i,e_k\rangle=\sum_k a_k\delta_{ik}=a_i.$$
 Geometrically, linear independence means each vector contributes something to the span of the vectors. So, if you have one vector, it spans a 1-dimensional subspace. If you add another vector to the set, you get a linearly dependent set if the span doesn't get any bigger. So, two vectors are linearly dependent if one is a multiple of the other. So, if you added a vector that was pointing in a different direction (and not the opposite direction), together they span a 2-dimensional subspace. In that case, they are said to be linearly independent. So, in general, if you have n vectors, they are linearly independent if throwing one of them out makes them span a smaller subspace. If you can throw one out without changing the span, they are linearly dependent. This picture is less accurate, but still helpful, for more general vector spaces, in which the "vectors" aren't exactly "arrows" pointing in space anymore. Functions are really a kind of vector because vectors are things that you can add together and multiply by scalars. So, it's the same idea for functions.
 Thank to all of you

 Similar discussions for: Linear independence Thread Forum Replies Precalculus Mathematics Homework 1 Calculus & Beyond Homework 3 Calculus & Beyond Homework 7 Calculus & Beyond Homework 2 Calculus & Beyond Homework 8