Prove matrix has all real eigenvalues

  • Context: MHB 
  • Thread starter Thread starter Jameson
  • Start date Start date
  • Tags Tags
    Eigenvalues Matrix
Click For Summary

Discussion Overview

The discussion revolves around proving that all eigenvalues of a symmetric matrix with real entries are real. Participants explore various approaches to this problem, including the use of complex inner-product spaces and the spectral theorem, while also addressing the prerequisites for understanding these concepts.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • Some participants propose using the property of symmetric matrices in complex inner-product spaces to show that eigenvalues must be real.
  • Others argue that the discussion should remain within the confines of real matrices, suggesting that complex analysis is not necessary for the course context.
  • A later reply questions the clarity of certain steps in the proof involving inner products and transposes, indicating a need for further explanation.
  • Some participants note that the characteristic polynomial of a matrix is of degree n and discuss the implications of this for the nature of its roots.
  • One participant suggests a more general statement regarding real eigenvalues and orthogonal eigenvectors, indicating that this would require a more complex proof.
  • Another participant mentions that every Hermitian matrix has real eigenvalues, linking this to the spectral theorem for normal matrices.

Areas of Agreement / Disagreement

Participants express differing views on the necessity of complex analysis in the proof, with some advocating for its inclusion and others preferring to avoid it. The discussion remains unresolved regarding the best approach to proving the statement about eigenvalues.

Contextual Notes

Some participants highlight the limitations of their understanding of inner-product spaces and the definitions involved, indicating that certain assumptions may not be universally accepted or understood.

Who May Find This Useful

This discussion may be useful for students and practitioners interested in linear algebra, particularly those studying properties of symmetric matrices and eigenvalues.

Jameson
Insights Author
Gold Member
MHB
Messages
4,533
Reaction score
13
Problem: Let $A$ be a $n \times n$ matrix with real entries. Prove that if $A$ is symmetric, that is $A = A^T$ then all eigenvalues of $A$ are real.

Solution: I'm definitely not seeing how to approach this problem. I know that to calculate the eigenvalues of a matrix I need to solve $\text{det }(A-\lambda I)=0$ and I have experience calculating them, but I've never seen commentary on whether the values will be real or complex. Any ideas to get started?
 
Last edited:
Physics news on Phys.org
Jameson said:
Problem: Let $A$ be a $n \times n$ matrix (with real entries). Prove that if $A$ is symmetric, that is $A = A^T$ then all eigenvalues of $A$ are real.
Think of $A$ as acting on the complex inner-product space $\mathbb{C}^n$ (whose inner product satisfies $\langle y,x\rangle = \overline{\langle x,y\rangle}$). If $\lambda$ is an eigenvalue of $A$, with eigenvector $x$, then $$\lambda\langle x,x\rangle = \langle Ax,x\rangle = \langle x,A^{\mathrm{\scriptsize T}}x\rangle = \langle x,Ax\rangle = \overline{\langle Ax,x\rangle} = \overline{\lambda}\langle x,x\rangle,$$ and so $\overline{\lambda} = \lambda$.
 
Thank you very much for the quick reply, Opalg! I don't believe this is the intended method to solve the problem though since any sort of complex analysis or complex theory isn't a prerequisite for the course. I'll take some time to read over your solution more and digest it, but am still looking perhaps for another potential method.
 
Jameson said:
Problem: Let $A$ be a $n \times n$ matrix. Prove that if $A$ is symmetric, that is $A = A^T$ then all eigenvalues of $A$ are real.

I suppse there is a typo. In order to be precise: "Let $A$ be a real $n \times n$ matrix ...". Oherwise, the result is false. Choose for example $A=\text{diag }(1,i).$ :)
 
Yes, you are very correct and my apologies for this error. :o I've fixed the OP now.
 
The way that Opalg approached the problem is the most common way that I've seen it done.

So, if you don't want to use complex numbers or spaces...

3vo4us.jpg


The other way to do this then would be to change the statement slightly and prove the following:

An $n\times n$ matrix $A$ has all real eigenvalues and $n$ orthogonal real eigenvectors if and only if $A$ is real symmetric.

The proof of the revised statement would require a lot more (and I do mean a lot more) work than the way Opalg presented.
 
If you're talking about roots of real polynomials, it's really hard to avoid a discussion of complex numbers.
 
My mistake everybody and thank you for reassurance. :) Seems I have some reading to do. I took linear algebra last semester but we didn't cover this method at all so it's completely new material. I'll try to work through Opalg's solution.
 
Hi Jameson!

What you are asking for is the proof of a general theorem called the "spectral theorem for real symmetric matrices".

Its proof is far from trivial, as you can see if you google for it, although Opalg's proof is quite elegant.
His proof is a bit concise though. I had to puzzle a bit to understand why the steps he's taking are valid.

First step is to realize that the characteristic polynomial you have, $\det(A-\lambda I)=0$, is a polynomial of degree n, meaning it has n (possible duplicate) roots if we allow complex numbers.
What's left to prove is that these eigenvalues are actually real numbers.
 
  • #10
If you start with complex numbers to begin with, you can actually say a bit more:

Every Hermetian matrix has real eigenvalues (Opalg's proof goes through as before).

As ILikeSerena has indicated, this is a consequence of the spectral theorem for normal matrices: every normal matrix is unitarily diagonalizable, and every unitarily diagonalizable matrix is normal

(normal matrices are those for which [math]AA^H = A^HA[/math]).

This is another instance of a statement about real numbers that makes more sense when you consider the wider context of complex numbers.

As far as I understand it (which is poorly) physicists prefer to work with complex vector spaces, because you can always restrict to the special case of real numbers, and complex numbers are in a sense "more complete". There's not much more difficulty involved in doing so, many of the results in linear algebra hold for an arbitrary field (there are some special cases where a field of characteristic 2 is inappropriate), and the usual definition of an inner product in a vector space over [math]\Bbb C[/math] resolves to a real-valued inner product when the scalars and coordinates are real.

What I'm trying to get across here is the idea that the "algebra" part of linear algebra, is based on the classical operations of addition, subtraction, multiplication and division (operations with matrices are "built-up" of operations of these kinds on individual entries) and a field is precisely the kind of algebraic object we can do those things IN.

For 99% of the examples one actually encounters in practice, the algebraic closure of the rational numbers would suffice (we rarely work with the full spectrum of transcendental numbers, a few important ones keep popping up). It is common practice when studying inner product spaces to assume the underlying field is a subfield of [math]\Bbb C[/math], and it is perhaps unfortunate that in introductory courses so much emphasis is given to [math]\Bbb R^n[/math], when you can do more without any extra effort.
 
  • #11
Opalg said:
Think of $A$ as acting on the complex inner-product space $\mathbb{C}^n$ (whose inner product satisfies $\langle y,x\rangle = \overline{\langle x,y\rangle}$). If $\lambda$ is an eigenvalue of $A$, with eigenvector $x$, then $$\lambda\langle x,x\rangle = \langle Ax,x\rangle = \langle x,A^{\mathrm{\scriptsize T}}x\rangle = \langle x,Ax\rangle = \overline{\langle Ax,x\rangle} = \overline{\lambda}\langle x,x\rangle,$$ and so $\overline{\lambda} = \lambda$.

Ok I've done some reading on inner-product spaces and feel comfortable with the basic axioms and some other conclusions from them. The steps I don't follow are $ \langle Ax,x\rangle = \langle x,A^{\mathrm{\scriptsize T}}x\rangle = \langle x,Ax\rangle$. The first two steps and the last three are clear but these three are not. Can someone explain?
 
  • #12
In some developments, one actually DEFINES the transpose of a square matrix [math]A [/math] as the matrix [math]B[/math] such that:

[math]\langle Ax,x \rangle = \langle x,Bx \rangle, \forall x [/math]

(this assumes a REAL inner product space).

However, we can follow this entry-by-entry:

[math]\langle Ax,x \rangle = \sum_i\left(\sum_j a_{ij}x_j\right)x_i[/math]

[math]= a_{11}x_1x_1 + a_{12}x_2x_1 + \cdots + a_{1n}x_nx_1 + a_{21}x_1x_2 + a_{22}x_2x_2 + \cdots + a_{2n}x_nx_2 +[/math]

[math]\cdots + a_{n1}x_1x_n + a_{n2}x_2x_n + \cdots + a_{nn}x_nx_n[/math]

[math] = x_1(a_{11}x_1 + a_{21}x_2 + \cdots + a_{n1}x_n) + x_2(a_{12}x_1 + a_{22}x_2 + \cdots + a_{n2}x_n) +[/math]

[math]\cdots + x_n(a_{1n}x_1 + a_{2n}x_2 + \cdots + a_{nn}x_n)[/math] <---plucking out the [math]x_j[/math] in the "middle position"

[math]= \sum_j x_j\left(\sum_i a_{ji}x_i \right) = \langle x,A^Tx \rangle[/math].

The equality [math]\langle A^Tx,x \rangle = \langle Ax,x \rangle [/math] follows because [math]A[/math] is symmetric, and thus [math]A^T = A[/math].

(Note: this assumes the [math]x_j[/math] are the coordinates of [math]x[/math] in some basis (the standard basis works well), and that the [math]a_{ij}[/math] are the entries of [math]A[/math] in that same basis).

(Note #2: I hope it's clear that when we rearrange the order of the [math]x_i[/math]'s and the [math]x_j[/math]'s in the complex case, the sesquilinearity of the complex inner product requires we use [math]\overline{a_{ji}}[/math] instead, leading to:

[math]\langle Ax,x \rangle = \langle x,A^Hx \rangle[/math]

This reduces to the symmetric case for the reals, since real numbers are self-conjugate).
 
Last edited:
  • #13
Alternatively:

The standard inner product is given by $\langle x,y \rangle = y^\dagger x$, whiere $\dagger$ denotes the conjugate transpose.
Since A is a real matrix, its conjugate transpose is the same as its transpose.
And since A is symmetric, its transpose is the same as A: $A^\dagger = A^T = A$.

So:
$$\langle Ax,x \rangle = x^\dagger(Ax) = (x^\dagger A)x = (A^\dagger x)^\dagger x = \langle x, A^\dagger x \rangle = \langle x, A^T x \rangle = \langle x, A x \rangle$$

Oh, and you may already now that in general $(MN)^T = N^TM^T$, which is the reason that we have here that $(x^\dagger A) = (A^\dagger x)^\dagger$.
 
Last edited:

Similar threads

  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 12 ·
Replies
12
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 10 ·
Replies
10
Views
2K
  • · Replies 14 ·
Replies
14
Views
4K
  • · Replies 2 ·
Replies
2
Views
2K