Linear independence of Coordinate vectors as columns & rows

Click For Summary
SUMMARY

The discussion centers on the linear independence of coordinate vectors represented as columns and rows in the context of the standard basis of ## M_2(\mathbb{R}) ##. The vectors ## \lbrack A \rbrack, \lbrack B \rbrack, \lbrack C \rbrack ## were initially thought to be linearly independent based on incorrect row operations, leading to confusion. Upon correcting the row operations, it was established that the vectors are indeed linearly dependent. The discussion also touches on the equivalence of representing vectors as rows or columns when checking for linear dependence.

PREREQUISITES
  • Understanding of linear algebra concepts, specifically linear independence and dependence.
  • Familiarity with matrix row operations and echelon forms.
  • Knowledge of the standard basis in vector spaces, particularly in ## M_2(\mathbb{R}) ##.
  • Basic understanding of the relationship between row space and column space of matrices.
NEXT STEPS
  • Study the properties of linear independence and dependence in vector spaces.
  • Learn about row echelon form and reduced row echelon form techniques in matrix algebra.
  • Explore the implications of the rank-nullity theorem in relation to linear transformations.
  • Investigate theorems regarding the equivalence of row and column representations in linear algebra.
USEFUL FOR

Students and professionals in mathematics, particularly those studying linear algebra, as well as educators seeking to clarify concepts of linear independence and matrix representations.

CGandC
Messages
326
Reaction score
34
Summary:: x

Question:
1608503464166.png


Book's Answer:
1608503489155.png


My attempt:

The coordinate vectors of the matrices w.r.t to the standard basis of ## M_2(\mathbb{R}) ## are:

##
\lbrack A \rbrack = \begin{bmatrix}1\\2\\-3\\4\\0\\1 \end{bmatrix} , \lbrack B \rbrack = \begin{bmatrix}1\\3\\-4\\6\\5\\4 \end{bmatrix} , \lbrack C \rbrack = \begin{bmatrix} 3\\8\\-11\\16\\10\\9 \end{bmatrix}
##
Putting these coordinate vectors in a matrix ( representing a homogeneous system of equations ):## \begin{bmatrix}
1 & 1 & 3 & | 0 \\
2 & 3 & 8 & | 0 \\
-3 & -4 & -11 & | 0 \\
4 & 6 & 16 & | 0 \\
0 & 5 & 10 & | 0 \\
1 & 4 & 9 & | 0 \\
\end{bmatrix} ##

After many row operations I get the matrix:

## \begin{bmatrix}
1 & 0 & 0 & | 0 \\
0 & 1 & 0 & | 0 \\
0 & 0 & 1 & | 0 \\
0 & 0 & 0 & | 0 \\
0 & 0 & 0 & | 0 \\
0 & 0 & 0 & | 0 \\
\end{bmatrix} ##

[Moderator's note: moved from a technical forum.]

Clearly we have 3 leading coefficients, three of them in the first 3 rows, therefore the coordinate vectors ## \lbrack A \rbrack , \lbrack B \rbrack , \lbrack C \rbrack ## are linearly independent, therefore the matrices ## A , B , C ## are linearly independent.

Why am I getting a contradiction to the real answer ( that ## A , B , C ## are linearly dependent )? How I could've a-priori known to represent the coordinate vectors as rows?

Is there a connection to column and row spaces?
 
Physics news on Phys.org
It doesn't matter if you represent them as rows or columns. You made a mistake in the work that you did, can you post it?
 
Turns out I made a mistake in the row operations ( even though I checked beforehand couple of times ), so I get the matrix:

##

\begin{bmatrix}

4 & 6 & 16 & | 0 \\

0 & 5 & 10 & | 0 \\

0 & 0 & 0 & | 0 \\

0 & 0 & 0 & | 0 \\

0 & 0 & 0 & | 0 \\

0 & 0 & 0 & | 0 \\

\end{bmatrix}

##
So the coordinate vectors are clearly linearly dependent.

I have another question: Is there some theorem stating that it won't matter to represent the vectors as columns or rows in order to check linear dependence?
 

Similar threads

  • · Replies 5 ·
Replies
5
Views
4K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 10 ·
Replies
10
Views
3K
Replies
3
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 16 ·
Replies
16
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K