Proof of linear independence and dependence

Click For Summary

Homework Help Overview

The discussion revolves around proving properties of linear independence and dependence in the context of matrix multiplication. Specifically, it examines two statements regarding matrices X and Y, focusing on the implications of linear dependence among the columns of Y and the resulting matrix product XY.

Discussion Character

  • Exploratory, Conceptual clarification, Mathematical reasoning, Assumption checking

Approaches and Questions Raised

  • Participants explore the implications of linear dependence in the columns of matrix Y and how it affects the columns of the product XY. There are attempts to clarify the mathematical notation and concepts, such as the dot product and linear combinations. Questions arise about the significance of certain assumptions and the correct interpretation of linear dependence.

Discussion Status

The discussion is ongoing, with participants providing insights and clarifications on the mathematical concepts involved. Some guidance has been offered regarding the structure of the proof, but there is still uncertainty among participants about specific details and terminology.

Contextual Notes

There is a noted lack of familiarity with certain concepts among participants, such as the dot product and linear combinations, which may affect their understanding of the problem. Additionally, there is some confusion regarding the assumptions made about the linear dependence of the columns in matrix Y.

andorrak
Messages
30
Reaction score
0
1. Homework Statement

There are two proofs:

Let X and Y be two matrices such that the product XY is defined. Show that if the columns of Y are linearly dependent, then so are the columns of the matrix XY.


Let X and Y be two matrices such that the product XY is defined. Show that if the columns of the matrix XY are linearly independent, then so are the columns of Y


2. Homework Equations
N/A


3. The Attempt at a Solution

Solution for first. i do not know

Solution for second one perhaps?

If XY are assumed to be the identity matrix. Thus we know I = (A^-1)(A)

Therefore, A=X and A^-1=Y. Then we know X^-1=Y. Then by the invertible matrix theorem, the equation Ax=0 has only the trivial solution and must be linearly independent?

Those are my two cents, can anyone help me?
 
Physics news on Phys.org
for the first part start with
[tex] Y = \begin{pmatrix} y_1 & .. & y_i .. & y_n \end{pmatrix} [/tex]

[tex]X = \begin{pmatrix} x_1^T \\ .. \\ x_j^T \\.. \\ x_m^T \end{pmatrix} [/tex]

now assume some y_i = a.y_r + b.ys (ie a linearly dependent column vectro for some r,s) and consider the action of the multiplication (each element in XY will correspond to a dot product between the row vector of X and the column vectros of Y)

so what is x^T dot ( a.y_r + b.ys )?
 
Last edited:
Um could you elaborate? I have no idea what you are saying. sorry
 
i re-wrote above for clarity & expanded a little
 
Sorry to sound nagging.

But I still have no idea what you are saying. Or what significance it brings. I am only on chapter 2 of the linear algebra book by Lay. I assumed it had something to do with the invertible matrix theorem
 
try expaninding this (for arbitrary vectors)

x^T dot ( a.y_r + b.ys )?
 
what is the dot product? if you could answer it systematically it would be very helpful
 
O that's just matrix multiplication. but i do not understand your notation, ie this.

try expaninding this (for arbitrary vectors)

x^T dot ( a.y_r + b.ys )?

so i assume x^T is the column for the X matrix. and r and s are the dependent vectors and the those dots between a.y and b.y i assume then are the dot products but where are the a's coming from? the entries of the x matrix?
 
  • #10
x_i^T is a row vector of X

say a column vector y_j can be written as a linear combination of other coulmn vectors y_r, y_s, for a,v constants

[tex]y_j = ay_r + a y_s[/tex]

[tex]x_i^T \bullet y_j = x_i^T \bullet (a y_r + b y_s)[/tex]
 
Last edited:
  • #11
so consider the jth column of XY
[tex] XY_j = \begin{pmatrix} x_1^T \bullet y_j\\ .. \\ x_i^T \bullet y_j\\.. \end{pmatrix} <br /> [/tex]
 
  • #12
OH i think i get it. because you assume all the columns of Y are linearly dependent you can show (which is what a and b represent) are t9eh coefficients when you have a linearly combination between the sets. Now when you multiply X^T with it, that is simply a scalar multiple of the one before so it IS STILL a linearly dependent vector.

Correct me if I am wrong. thanks!
 
  • #13
pretty much, however i would say all you need to assume is there is a linearly dependent column in Y, so you may want to generalise a bit
 
  • #14
Yea but the question assumes every column in Y is linearly dependent so I am set. thanks again !
 
  • #15
andorrak said:
Yea but the question assumes every column in Y is linearly dependent so I am set. thanks again !

no it doesn't that statement doesn't even make sense

it says "the columns of Y are linearly dependent" this means at least one cloumn can be written as a linear combination of the others
 

Similar threads

  • · Replies 30 ·
2
Replies
30
Views
2K
Replies
1
Views
2K
Replies
2
Views
1K
Replies
4
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 3 ·
Replies
3
Views
1K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K