Transforming a matrix to orthogonal one

  • Context: Graduate 
  • Thread starter Thread starter onako
  • Start date Start date
  • Tags Tags
    Matrix Orthogonal
Click For Summary

Discussion Overview

The discussion revolves around the transformation of a matrix with linearly independent columns into an orthogonal matrix. Participants explore the conditions under which a matrix A can be found such that the product of the original matrix X and A results in an orthonormal matrix Y, satisfying the equation Y^TY=I. The conversation includes theoretical considerations and mathematical reasoning related to the Gram-Schmidt process and properties of matrices.

Discussion Character

  • Technical explanation
  • Mathematical reasoning
  • Debate/contested

Main Points Raised

  • One participant proposes that if a matrix X has p linearly independent columns, there exists a matrix A that can transform X into an orthonormal matrix Y.
  • Another participant confirms this by suggesting the use of the Gram-Schmidt orthogonalization process to construct an orthonormal basis from the columns of X.
  • A participant raises a clarification regarding the dimensionality of the basis in R^n and questions how to calculate matrix A a priori.
  • Another participant provides an alternative explanation involving the properties of the matrix X^*X, asserting that it is invertible and positive definite, leading to the derivation of matrix A as the inverse of a square root of X^*X.
  • A participant expresses appreciation for the detailed explanation provided regarding the properties of the matrices involved.

Areas of Agreement / Disagreement

Participants generally agree on the existence of a matrix A that can transform X into an orthonormal matrix Y, but there are differing explanations regarding the underlying reasons and methods to derive A. The discussion remains unresolved on the specifics of calculating A a priori.

Contextual Notes

Participants discuss the implications of linear independence and the properties of matrices without resolving the mathematical steps involved in calculating A. The conversation reflects varying approaches to the problem, highlighting the complexity of the topic.

onako
Messages
86
Reaction score
0
Suppose a matrix X of size n x p is given, n>p, with p linearly independent columns. Can it be guaranteed that there exists a matrix A of size p x p that converts columns of X to orthonormal columns. In other words, is there an A, such that Y=XA, and Y^TY=I, where I is an p x p identity matrix.
 
Physics news on Phys.org
Yes. Since the columns of X are independent, they for a basis for Rn. We can then use the "Gram-Schmidt orthogonalization" process to construct an orthonormal basis from them. A will be the "change of basis" matrix that changes representation of a vector in the original basis to representation in the orthonormal basis.
 
Thanks. Just one note: I suppose you've taken into account that there are p columns in X (which is an n x p matrix). If I'm not wrong, only n linearly independent columns of dimensionality R^n define a basis in R^n.

So, given an input X, with linearly independent columns, such columns could be transformed by GS processing to yield Y, such that Y^TY=I, and there exists A, such that Y=XA. How could one calculate such A a priori?
 
Yes Onako, it is true, but for a different reason, than stated by HallsofIvy.

If the columns of X are linearly independent, X^*X is an invertible p\times p matrix (here X^* is the Hermitian conjugate of X, i.e. the conjugate transpose of X ; if X is real then X^*=X^T (I use X^* only because what I say works for complex matrices as well).

Matrix X^*X is positive semidefinite for all X , and since X^*X is invertible, X^*X is positive definite (all eigenvalues are positive). Since the matrix X^*X is Hermitian (symmetric if X is real), it is can be diagonalized, i.e. it can be represented as a diagonal matrix in some orthonormal basis, or equivalently, it can be written as X^*X =U^* D U, where U is a unitary matrix (U^{-1}=U^*) and D is a diagonal matrix with eigenvalues of X^*X on the diagonal.

We can take a square root of X^*X, namely B = U^* D^{1/2} U, where D^{1/2} is obtained by taking square roots of diagonal entries of D (recall that D is a diagonal matrix). Then B^*=B, and X^*X = B^2, and A=B^{-1} is the matrix you want.

Indeed, if Y=XA, then Y^*Y = A^* X^*X A = A^* B^2 A =A B^2 A =I.
 
Thank your for such a good explanation.
 

Similar threads

  • · Replies 14 ·
Replies
14
Views
4K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 14 ·
Replies
14
Views
4K
  • · Replies 33 ·
2
Replies
33
Views
3K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 14 ·
Replies
14
Views
3K
  • · Replies 1 ·
Replies
1
Views
4K
  • · Replies 27 ·
Replies
27
Views
3K