Linearly independent vectors and span

Click For Summary

Homework Help Overview

The discussion revolves around the properties of linearly independent vectors within a vector space, specifically addressing the question of whether a subset of these vectors can span the entire space when one vector is removed.

Discussion Character

  • Conceptual clarification, Assumption checking

Approaches and Questions Raised

  • Participants explore the definition of linear independence and question the implications of removing a vector from a linearly independent set. There is a suggestion to consider how one vector might be expressed as a linear combination of others, prompting further examination of the definitions involved.

Discussion Status

Participants are actively engaging with the definitions and implications of linear independence and span. Some guidance has been offered regarding the structure of a proof, including steps to clarify definitions and assumptions. Multiple interpretations of the problem are being explored.

Contextual Notes

There is an emphasis on precise language and definitions, as well as the context of working within an n-dimensional vector space, which may influence the understanding of the problem.

jasoqueso
Messages
1
Reaction score
0
So the question is...

Let v1, v2,...,,vn be linearly independent vectors in a vector space V. Show that v2,...,vn cannot span V.

I honestly have found myself completely lost lately and I suck at writing proofs.

So this is what I see,

v1, v2,..., vn is linearly independent iff c1v1+c2v2+ ... + cnvn = 0 and c1=c2=...=cn= 0

however I don't see why taking away v1 would make it not span V anymore.
 
Physics news on Phys.org
jasoqueso said:
So the question is...

Let v1, v2,...,,vn be linearly independent vectors in a vector space V. Show that v2,...,vn cannot span V.

I honestly have found myself completely lost lately and I suck at writing proofs.

So this is what I see,

v1, v2,..., vn is linearly independent iff c1v1+c2v2+ ... + cnvn = 0 and c1=c2=...=cn= 0

however I don't see why taking away v1 would make it not span V anymore.

Consider this. How do you write v1 as a linear combination of v2, ..., vn?

Or for a concrete example, in R^3. Take three vectors at right angles called x, y, and z. Can you write x as a linear combination of y and z? Obviously not, but go back to your definition of "span" and "linearly independent" and figure out why!
 
Do the following:

1. Re-state the definition of linear independence carefully. (Your language was unclear.)
2. Assume that you can express v1 as a linear combination of the other vectors.
3. Show that this will contradict linear independence.
4. The conclusion is that you cannot express v1 using the other vectors.

It's very important to know your definitions precisely!
 
In an n-dimensional vector space, a linearly independent set of n vectors is a basis for that space. Also, in an n-dimensional vector space, every basis for that space has the same amount of vectors. A basis is a linearly independent set that spans the space. This makes the rest of the proof easy.
 
Last edited:

Similar threads

Replies
17
Views
3K
Replies
4
Views
2K
Replies
15
Views
2K
Replies
1
Views
4K
  • · Replies 2 ·
Replies
2
Views
3K
Replies
12
Views
2K
  • · Replies 12 ·
Replies
12
Views
3K
Replies
1
Views
5K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K