Zero eigenvalues or eigenvectors

nomadreid
Gold Member
Messages
1,748
Reaction score
243
I have a bit of problem with zero eigenvectors and zero eigenvalues. On one hand, there seems to be nothing in the definition that forbids them, and they even seem necessary to allow because an eigenvalue can serve as a measurement and zero can be a measurement, and if there is a zero eigenvalue then it will be a term in a diagonalized matrix, so that one has a zero eigenvector as well (a column vector of the diagonal matrix with the zero eigenvalue). So far, so good. But on the other hand, if the zero eigenvector is allowed, then every value in the field would be an eigenvalue, hence making it a bit trivial, no?
 
Physics news on Phys.org
Eigenvectors are non-zero by definition.

An eigenvalue of zero on the other hand is fine. If you have a zero column in your diagonal matrix, you have to chose a non-zero value for the entry which gets multiplied by the zero eigenvalue in order to get a proper eigenvector.
 
  • Like
Likes nomadreid
I have never seen a zero eigenvector, but zero eigenvectors come up in the theory of vibrations often enough. In that context, it means that the system is positive semi-definite, rather than positive definite. In physical terms, it means that the stiffness matrix is singular and that rigid body motion is possible.
 
Thanks, kith and Dr. D.
kith: that clears it up nicely, thanks.
Dr. D. Good to know: except I presume you had a typo, in that you meant , instead of
Dr.D said:
but zero eigenvectors come up in the theory of vibrations
that "but zero eigenvalues come up in the theory of vibrations"
 
Note that, generally speaking, zero eigenvalues are nothing special. Given a matrix A with an eigenvector x,
$$
A x = \lambda x
$$
I can construct a new matrix ##B = A - \lambda I##, where ##I## is the identity matrix, such that
$$
B x = 0 x
$$
This is called eigenvalue shifting, and is used in numerical methods for eigenequations, in order to speed up convergence.
 
  • Like
Likes nomadreid
There is a slight ambiguity in the definition of "eigenvector". Most textbooks define "eigenvalue" for a linear operator, A, as a number, \lambda such that there exist a non-zero vector v with Av= \lambda v and then define an "eigenvector" corresponding to eigenvalue \lambda as such a non-zero vector. But some define eigenvalue in that way and then define "eigenvector" as any vector, v, satisfying Av= \lambda v. I prefer that- it allows one to say things like "any multiple of an eigenvector is an eigenvector" without having to say "except 0" and "the set of all eigenvectors corresponding to eigenvalue \lambda form a vector space" without having to say "with the 0 vector added".
 
Last edited by a moderator:
  • Like
Likes nomadreid
Nomadreid, yes, you are correct. I was thinking one thing and typing another.
 
  • Like
Likes nomadreid
Back
Top