Non-square matrices and inverses

  • Context: Undergrad 
  • Thread starter Thread starter Bipolarity
  • Start date Start date
  • Tags Tags
    Matrices
Click For Summary

Discussion Overview

The discussion centers on the properties of inverses for non-square matrices, specifically whether inverses can exist for such matrices and under what conditions. Participants explore the implications of matrix dimensions, linear independence, and the concept of pseudoinverses, while examining the limitations of non-square matrices in relation to their inverses.

Discussion Character

  • Debate/contested
  • Technical explanation
  • Mathematical reasoning

Main Points Raised

  • One participant questions whether inverses for non-square matrices do not exist because no (m by n) matrix A can have an (n by m) matrix B such that both AB = I and BA = I.
  • Another participant explains that if n > m, AB will be an (m by m) matrix and BA will be an (n by n) matrix, thus they cannot both equal the identity matrix I.
  • It is noted that if the m columns of A are linearly independent, several matrices B can exist such that AB = I(m by m), but no matrix B can satisfy BA = I(n by n).
  • Conversely, if n < m, left inverses can exist such that BA = I(n by n), but no right inverses can satisfy AB = I(m by m).
  • A participant introduces the concept of the pseudoinverse as a potential solution for finding a matrix A such that AB = I(m by m) and BA = I(n by n).
  • Another participant provides a specific example illustrating the impossibility of having both AB = I and BA = I unless m = n and A is an invertible square matrix.
  • Concerns are raised about the generalizability of the provided examples and proofs to all (m by n) matrices.
  • A participant discusses the implications of dimensions on linear transformations and the existence of inverses, emphasizing that if m > n, the mapping is not onto, and if n > m, the mapping is not one-to-one.

Areas of Agreement / Disagreement

Participants express differing views on the existence of inverses for non-square matrices, with some asserting that certain conditions allow for right or left inverses, while others maintain that both cannot exist simultaneously. The discussion remains unresolved regarding the general case of (m by n) matrices.

Contextual Notes

Participants reference the rank of matrices and its implications for the existence of inverses, but the discussion does not resolve the mathematical steps or assumptions involved in these arguments.

Bipolarity
Messages
773
Reaction score
2
I now know that inverses are only defined for square matrices. My question is: is this because inverses for non-square matrices do not exist, i.e. there is no (m by n) matrix A for which there exists an (n by m) matrix B such that both AB = I and BA = I is true?

Or is it just done for convenience? In this cases, can we indeed find a (m by n) matrix A for which there exists an (n by m) matrix B such that both AB = I and BA = I is true?

Thanks!

BiP
 
Physics news on Phys.org
AB will be an (m by m) matrix, and BA will be (n by n), so they can't both be equal to the same matrix I.

If n>m and the m columns of A are linearly independent, you can find several matrices B with AB = I(m by m), and these matrices B are "right inverses" of A. You can't, however, find a matrix B with BA = I(n by n), for the following reason:

Let rank(M) stand for the maximum number of linearly independent rows or columns of M. (The numbers can be proven to be equal.) Then rank(M) ≤ min{ columns of M, rows of M } and rank(AB) ≤ min{ rank(A), rank(B) }.

In your case, rank(AB) ≤ m, but rank(I(n by n)) = n.

(If n<m, you can find several matrices "left inverse" matrices B with BA = I(n by n), but no matrix B with AB = I(m by m).)
 
Michael Redei said:
AB will be an (m by m) matrix, and BA will be (n by n), so they can't both be equal to the same matrix I.

If n>m and the m columns of A are linearly independent, you can find several matrices B with AB = I(m by m), and these matrices B are "right inverses" of A. You can't, however, find a matrix B with BA = I(n by n), for the following reason:

Let rank(M) stand for the maximum number of linearly independent rows or columns of M. (The numbers can be proven to be equal.) Then rank(M) ≤ min{ columns of M, rows of M } and rank(AB) ≤ min{ rank(A), rank(B) }.

In your case, rank(AB) ≤ m, but rank(I(n by n)) = n.

(If n<m, you can find several matrices "left inverse" matrices B with BA = I(n by n), but no matrix B with AB = I(m by m).)

I see. What if we want the (m by n) matrix A so that AB = I where I is (m by m) and BA = I where I is (n by n).

Does such a matrix exist?

BiP
 
Robert1986 said:
Yes, it is called the pseudoinverse: http://en.wikipedia.org/wiki/Moore–Penrose_pseudoinverse

Thanks! Could you provide an example of such a matrix, i.e. an (m by n) matrix A such that there is an (n by m) matrix B for which AB = I_{m} and BA = I_{n} where the subscript of I denotes its size.

BiP
 
Bipolarity said:
I see. What if we want the (m by n) matrix A so that AB = I where I is (m by m) and BA = I where I is (n by n).

You can't have both, unless m=n and A is an invertible square matrix. Consider this simple example:
<br /> \left( \begin{array}{cc} 1 &amp; 0 \\ 0 &amp; 1 \\ 1 &amp; 1 \end{array} \right) \cdot <br /> \left( \begin{array}{ccc} a &amp; b &amp; c \\ d &amp; e &amp; f \end{array} \right) =<br /> \left(\begin{array}{ccc} a &amp; b &amp; c \\ d &amp; e &amp; f \\ a + d &amp; b + e &amp; c + f \end{array} \right) =<br /> \left(\begin{array}{ccc} 1 &amp; 0 &amp; 0 \\ 0 &amp; 1 &amp; 0 \\ 0 &amp; 0 &amp; 1 \end{array} \right),<br />i.e. both c=f=0 and c+f=1, which is impossible. But you can have
<br /> \left( \begin{array}{ccc} a &amp; b &amp; c \\ d &amp; e &amp; f \end{array} \right) \cdot<br /> \left( \begin{array}{cc} 1 &amp; 0 \\ 0 &amp; 1 \\ 1 &amp; 1 \end{array} \right) = <br /> \left(\begin{array}{cc} a + c &amp; b + c \\ d + f &amp; e + f \end{array} \right) =<br /> \left(\begin{array}{cc} 1 &amp; 0 \\ 0 &amp; 1 \end{array} \right)<br />if you set
<br /> \left( \begin{array}{ccc} a &amp; b &amp; c \\ d &amp; e &amp; f \end{array} \right) =<br /> \left( \begin{array}{ccc} a &amp; (a-1) &amp; (1-a) \\ d &amp; (d+1) &amp; -d \end{array} \right).<br />
 
Thanks! But is this true for general matrices of size (m by n)? How would one prove it in the general case? It seems really difficult to extrapolate this proof to such general forms

BiP
 
Please have another look at what I wrote about ranks.
 
More generally, a "m by n" matrix, A, with m columns and n rows, represents a linear transformation from an n dimensional vector space, U, to an m dimensional vector space, V. If m> n, then A maps U into an n dimensional subspace of V. Since n< m, the mapping is not "onto"- there exist vectors, v, in V that are not in the that subspace. That means there was no "u" in U such that Au= v and so no vector for "A^{-1}" to map v back to.

If, on the other hand, n> , then A maps U into all of V, but NOT "one to one". That is, given v in V, there must exist at least 2 (actually and infinite number) vectors, u1 and u2, such that A(u1)= A(u2)= v. A single inverse matrix cannot map v back to both u1 and u2.
 

Similar threads

  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 12 ·
Replies
12
Views
2K
Replies
13
Views
4K
  • · Replies 34 ·
2
Replies
34
Views
3K