Any square matrix can be expressed as the sum of anti/symmetric matrices

Click For Summary

Discussion Overview

The discussion revolves around the proposition that any square matrix can be expressed as the sum of symmetric and anti-symmetric matrices. Participants explore the definitions and properties of symmetric and anti-symmetric matrices, as well as the implications of this proposition in various contexts.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • One participant defines matrices X and Y based on the entries of a square matrix A, concluding that A can be expressed as the sum of X (symmetric) and Y (anti-symmetric).
  • Another participant suggests a minimalist formulation of the proposition, indicating that 2A can be expressed as the sum of A and its transpose, and the difference of A and its transpose.
  • Concerns are raised about the need to prove that the sum of a matrix and its transpose is symmetric, and the difference is anti-symmetric, with some arguing that this is an obvious proposition that could be proven succinctly.
  • One participant expresses uncertainty about the necessity of defining "square" and suggests focusing on the more challenging aspects of the proof first.
  • Another participant agrees that proving the properties of the sum and difference of matrices is straightforward, questioning the need for a formal proof.
  • There is a suggestion to consider the implications in characteristic 2, where the claim may not hold, indicating that skew symmetric matrices are a subset of symmetric matrices in this context.

Areas of Agreement / Disagreement

Participants express differing views on the necessity of proving certain properties related to symmetric and anti-symmetric matrices. While some find the propositions obvious, others suggest that formal proofs may still be warranted. Additionally, there is a recognition of potential limitations in the context of characteristic 2, indicating a lack of consensus on the universality of the initial claim.

Contextual Notes

Some participants highlight the importance of assumptions and definitions, particularly regarding the term "square" and the implications of characteristic 2 on the validity of the claim.

Eclair_de_XII
Messages
1,082
Reaction score
91
TL;DR
es. Show that any square matrix can be expressed as the sum of a symmetric matrix and an anti-symmetric matrix of the same size.

Important: I would like critique on the style of my proof more than its content. That being said, I am more than open to critique on the latter, as well. In my academic career, I've often been told that my proofs are very long and needlessly excessive. This is the very first proof-based exercise in my linear algebra book, and it took about a page for me to write.
Let ##A## be a matrix of size ##(n,n)##. Denote the entry in the i-th row and the j-th column of ##A## by ##a_{ij}##, for some ##i,j\in\mathbb{N}##. For brevity, we call ##a_{ij}## entry ##(i,j)## of ##A##.

Define the matrix ##X## to be of size ##(n,n)##, and denote entry ##(i,j)## of ##X## as ##x_{ij}##, where ##x_{ij}=\frac{1}{2}\left(a_{ij}+a_{ji}\right)##. We note that ##x_{ji}=\frac{1}{2}\left(a_{ji}+a_{ij}\right)=\frac{1}{2}\left(a_{ij}+a_{ji}\right)=x_{ij}##. Since matrices are uniquely identified by their entries, we conclude from this string of equalities that ##X^T=X##.

Define the matrix ##Y## to be of size ##(n,n)##. Denote entry ##(i,j)## of ##Y## as ##y_{ij}##, where ##y_{ij}=\frac{1}{2}\left(a_{ij}-a_{ji}\right)##. We note that ##y_{ji}=\frac{1}{2}\left(a_{ji}-a_{ij}\right)=-\frac{1}{2}\left(a_{ij}-a_{ji}\right)=-y_{ij}##. Since matrices are uniquely identified by their entries, we conclude from this string of equalities that ##Y^T=Y##.

\begin{eqnarray}
x_{ij}+y_{ij}&=&\frac{1}{2}\left(a_{ij}+a_{ji}\right)+\frac{1}{2}\left(a_{ij}-a_{ji}\right)\\
&=&\frac{1}{2}a_{ij}+\frac{1}{2}a_{ji}+\frac{1}{2}a_{ij}-\frac{1}{2}a_{ji}\\
&=&\left(\frac{1}{2}a_{ij}+\frac{1}{2}a_{ij}\right)+\left(\frac{1}{2}a_{ji}-\frac{1}{2}a_{ji}\right)\\
&=&a_{ij}+0\\
&=&a_{ij}
\end{eqnarray}

Since matrices are uniquely identified by their entries, we conclude that ##A=X+Y##.
 
Physics news on Phys.org
I guess the absolute minimalist version of this is
Note that for a square matrix A $$2\mathbf A= [\mathbf A+{\mathbf A}^t]+ [ \mathbf A-{\mathbf A}^t] $$
 
Wouldn't you need to prove that the sum of any matrix and its transpose is symmetric, and the difference of any matrix and its transpose is anti-symmetric? Or is that just a proposition that is too obvious to prove --- which is to say, it could be proven in just a handful of lines of algebra?
 
I would not feel the need but I am not a mathematician. Do you need to define what "square" means?
I understand your angst and it used to drive me crazy..."where do I start??""what do I assume?"
I suggest that you figure out the really hard part first and write it down. Then anything else is is just the decoration.. maybe you need it
I would appreciate other opinions.
 
  • Like
Likes   Reactions: Keith_McClary
Eclair_de_XII said:
Wouldn't you need to prove that the sum of any matrix and its transpose is symmetric, and the difference of any matrix and its transpose is anti-symmetric? Or is that just a proposition that is too obvious to prove --- which is to say, it could be proven in just a handful of lines of algebra?
It is too obvious to prove. What are ##\left(A^\tau\right)^\tau## and ##(A+B)^\tau ?##
 
Oh. So set ##B=\pm A^T##.
 
Eclair_de_XII said:
Oh. So set ##B=A^T##.
Yes. You use linearity of transposition and that it is inverse to itself. If you transpose a product, then you must change the order of the factors, but this isn't used here.
 
  • Like
Likes   Reactions: Frabjous
hutchphd said:
I guess the absolute minimalist version of this is
Note that for a square matrix A $$2\mathbf A= [\mathbf A+{\mathbf A}^t]+ [ \mathbf A-{\mathbf A}^t] $$
This is how I'd write it, with a footnote saying that in the case of characteristic 2, this isn't meaningful and OP's claim isn't true. (Skew symmetric matrices are a subset of symmetric matrices in characteristic 2, so by a dimension argument it can't be true or just exhibiting a single non-symmetric matrix gives a counterexample.)
 

Similar threads

  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 31 ·
2
Replies
31
Views
2K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
Replies
1
Views
1K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 27 ·
Replies
27
Views
2K