[Linear Algebra] Maximal linear independent subset

Click For Summary

Homework Help Overview

The discussion revolves around finding a maximal linearly independent subset of given sets of vectors in different contexts: one in \(\mathbb{R}^4\) and the other in a polynomial space. The original poster expresses confusion regarding the representation of vectors as rows versus columns in a matrix and how this affects the determination of linear independence.

Discussion Character

  • Exploratory, Assumption checking, Mathematical reasoning

Approaches and Questions Raised

  • The original poster attempts to understand the mechanics of representing vectors in matrix form, questioning whether to use rows or columns. They share their previous experience with a different set of vectors and the results obtained from both representations.
  • Some participants question the correctness of the matrix representation and suggest that the rank should remain consistent regardless of the orientation of the vectors.
  • Others clarify that row operations affect the linear combinations differently depending on whether vectors are treated as rows or columns.

Discussion Status

The discussion is ongoing, with participants providing clarifications and addressing the original poster's confusion. There is an exploration of the implications of matrix representation on the results of linear independence tests, but no consensus has been reached on the best approach.

Contextual Notes

Participants note potential errors in matrix entries and emphasize the importance of consistent representation in linear algebra. The original poster's confusion stems from differing outcomes based on how vectors are arranged in the matrix.

iJake
Messages
41
Reaction score
0

Homework Statement



In the follow cases find a maximal linearly independent subset of set ##A##:
(a) ##A = \{(1,0,1,0),(1,1,1,1),(0,1,0,1),(2,0,-1,)\} \in \mathbb{R}^4##
(b) ##A = \{x^2, x^2-x+1, 2x-2, 3\} \in \mathbb{k}[x]##

The Attempt at a Solution



The first part of the exercise is trivial, as it is easy to observe that the second vector is a linear combination of the first and third vectors. My question is one of mechanics. Should I be writing the elements of each vector as the rows or columns of a matrix?

In a previous exercise, I had to determine whether or not a vector space was spanned by a set of vectors. In that case, it was:

Determine if ## V = \mathbb{k}[x]_3## is spanned by ##A = \{1, 1+x^2, 1-x+x^2+x^3, 4-x+2x^2+x^3\}##

I reordered the elements and presented them as the columns of a matrix:

##
\begin{matrix}
0 & 0 & 1 & 1\\
0 & 1 & 1 & 2\\
0 & 0 & -1 & -1\\
1 & 1 & 1 & 4\\
\end{matrix}
##

I did row operations (switch R4 with R1, then with R3, add R3 to R4) and found a row of 0s, thus telling me the rank of the matrix is 3 and that A does not span V.

However, for my current problem, I wrote the matrix using the vector elements as rows, not columns. When I write them as columns I do not reach the same conclusion, which confuses me.

##
\begin{matrix}
1 & 0 & 1 & 0\\
1 & 1 & 1 & 1\\
0 & 1 & 0 & 1\\
2 & 0 & -1 & 0\\
\end{matrix}
##

Here I clearly observe that the second row is the sum of the first and third rows. However, if I write it where the vector elements are columns, I see something else:

##
\begin{matrix}
1 & 1 & 0 & 2\\
0 & 1 & 1 & 0\\
1 & 1 & 0 & -1\\
0 & 1 & 1 & 0\\
\end{matrix}
##

Can someone explain the difference to me? Thanks.
 
Last edited:
Physics news on Phys.org
It might be because the last matrix has an error in it. The bottom left element should be 0, not 1.

If all you are trying to do is test whether n vectors in a n-dimensional space are independent, by calculating the rank of a matrix, you should get the same result regardless of whether you make the vectors rows or columns. I think it is more usual to make the vectors columns, but that's just convention.
 
Sorry, that was just a typo. It's fixed now, and I find that R4 - R2 gives me a row of 0s and that the rank of that matrix is 3. Is it irrelevant that the linearly dependent vector corresponds to the second column of the matrix, in this case?
 
iJake said:

Homework Statement



In the follow cases find a maximal linearly independent subset of set ##A##:
(a) ##A = \{(1,0,1,0),(1,1,1,1),(0,1,0,1),(2,0,-1,)\} \in \mathbb{R}^4##
(b) ##A = \{x^2, x^2-x+1, 2x-2, 3\} \in \mathbb{k}[x]##

The Attempt at a Solution



The first part of the exercise is trivial, as it is easy to observe that the second vector is a linear combination of the first and third vectors. My question is one of mechanics. Should I be writing the elements of each vector as the rows or columns of a matrix?

In a previous exercise, I had to determine whether or not a vector space was spanned by a set of vectors. In that case, it was:

Determine if ## V = \mathbb{k}[x]_3## is spanned by ##A = \{1, 1+x^2, 1-x+x^2+x^3, 4-x+2x^2+x^3\}##

I reordered the elements and presented them as the columns of a matrix:

##
\begin{matrix}
0 & 0 & 1 & 1\\
0 & 1 & 1 & 2\\
0 & 0 & -1 & -1\\
1 & 1 & 1 & 4\\
\end{matrix}
##

I did row operations (switch R4 with R1, then with R3, add R3 to R4) and found a row of 0s, thus telling me the rank of the matrix is 3 and that A does not span V.

However, for my current problem, I wrote the matrix using the vector elements as rows, not columns. When I write them as columns I do not reach the same conclusion, which confuses me.

##
\begin{matrix}
1 & 0 & 1 & 0\\
1 & 1 & 1 & 1\\
0 & 1 & 0 & 1\\
2 & 0 & -1 & 0\\
\end{matrix}
##

Here I clearly observe that the second row is the sum of the first and third rows. However, if I write it where the vector elements are columns, I see something else:

##
\begin{matrix}
1 & 1 & 0 & 2\\
0 & 1 & 1 & 0\\
1 & 1 & 0 & -1\\
0 & 1 & 1 & 0\\
\end{matrix}
##

Can someone explain the difference to me? Thanks.
Both
$$A = \begin{bmatrix}
1 & 0 & 1 & 0\\
1 & 1 & 1 & 1\\
0 & 1 & 0 & 1\\
2 & 0 & -1 & 0\\
\end{bmatrix}
$$
and its transpose
$$B = \begin{bmatrix}
1 & 1 & 0 & 2\\
0 & 1 & 1 & 0\\
1 & 1 & 0 & -1\\
0 & 1 & 1 & 0\\
\end{bmatrix}
$$
have the same row and column ranks---3 in this case.

If you perform row-operations on ##A## you will produce a new matrix whose rows are linear combinations of the row-vectors ##\vec{a}_i## in ##A##, so will be manufacturing directly some linear combination ##c_1 \vec{a}_1 + c_2 \vec{a}_2 + c_3 \vec{a}_3 + c_4 \vec{a}_4## that is the zero (row) vector; that will be the "all-zero" row of the transformed matrix. When you perform row operations on ##B## that is not what you are doing: the new columns are not linear combinations of the old columns, and row operations that produced a zero-row in ##A## will not necessarily produce a zero row in ##B##---although, of course, some other row operations will do so. To find out what linear combinations of the column vectors ##\vec{b}_i## in ##B## are needed to get the zero (column) vector, we need to find ##k_i## that give ##k_1 \vec{b}_1 + k_2 \vec{b}_2 + k_3 +\vec{b}_3 + k_4 \vec{b}_4 = \vec{0}##. That can be done by solving the equation system
$$B \vec{K} = \vec{0},$$
where
$$\vec{K} = \pmatrix{k_1\\ k_2 \\ k_3 \\ k_4}.$$
Row operations on ##B## are just helping you to solve that system of equations.
 

Similar threads

  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 8 ·
Replies
8
Views
1K
Replies
9
Views
2K
  • · Replies 10 ·
Replies
10
Views
3K
Replies
4
Views
2K
  • · Replies 30 ·
2
Replies
30
Views
2K
Replies
15
Views
2K
Replies
1
Views
2K
Replies
8
Views
2K