Motivation behind eigenvalue and eigenvector

Click For Summary
Eigenvectors are defined as non-zero vectors that, when multiplied by a square matrix, preserve their direction, represented mathematically as A.v = λ.v. The discussion highlights the practical applications of eigenvectors, particularly in Principal Component Analysis (PCA), where they help identify uncorrelated variables from correlated data sets. The conversation also touches on the significance of diagonalizing matrices, as eigenvectors and eigenvalues facilitate this process, simplifying complex matrices into more manageable forms. Additionally, there is a debate about the approach to finding eigenvectors versus transformations, with some suggesting it might be more intuitive to find a transformation for a given vector instead. Ultimately, understanding eigenvectors and their applications is crucial in various mathematical and engineering contexts.
Avichal
Messages
294
Reaction score
0
An eigenvector is defined as a non-zero vector 'v' such that A.v = λ.v

I don't understand the motive behind this. We are trying to find a vector that when multiplied by a given square matrix preserves the direction of the vector.

Shouldn't the motive be the opposite i.e. finding the matrix A given the vector v?

I suppose eigenvector was defined this way with some application in mind
 
Physics news on Phys.org
Avichal said:
I suppose eigenvector was defined this way with some application in mind

There are many applications. For me, one obvious application is to find the set of uncorrelated variables in large data sets. Such data sets are usually initially described by correlated variables and the object of the analysis is to transform the data into its principal components (called Principal Component Analysis or PCA). The principal components are the eigenvectors of the covariance matrix describing the relations among the initial set of variables. Each eigenvector represents one new variable. These new variables are independent of each other (orthogonal). The eigenvalues of the principal components correspond to the variance of each component.
 
Last edited:
Avichal said:
An eigenvector is defined as a non-zero vector 'v' such that A.v = λ.v

I don't understand the motive behind this. We are trying to find a vector that when multiplied by a given square matrix preserves the direction of the vector.

It's not just "a" vector, it could be several.

One path to understanding is to think about the convenience that a "change of coordinates" brings to many problems. If M is a matrix that represents something in a given set of coordinates, what happens to M when you change coordinates? (If you look into this you'll find that a linear change of coordinates amounts to the multiplication A^{-1} M A for some matrix A )

If M is a complicated matrix, and you want to change coordinates to make it simple, what is the "simplest" type of matrix to deal with? I think a diagonal matrix is simplest. You can't always change coordinates to revise the information in a matrix M so it is displayed as a diagonal matrix, but in many important cases you can. Suppose you do get a diagonal matrix. The eigenvectors of a diagonal matrix are simple to find, aren't they? The eigen values of these eigenvectors are just the numbers on the diagonal. If you only have to deal with diagonal matrices then eigenvectors and eigen values are obviously important concepts - although they are triival to find.

If we use a change of coordinates that doesn't change any vectorial properties of the information in M then the eigenvectors of the diagonalized version of the information should be the eigenvectors of M expressed in a different system of coordinates. If you don't know how to diagonalize M by changing coordinates then finding the eigenvectors of M can give you a hint about how to do it.

For matrices that can't be diagonalized by a change of coordinates, it turns out that they can be changed to a somewhat simple form consisting of "blocks" of numbers calld "Jordan blocks". Eigenvalues also play a role in the facts about that.
 
Consider the differential equation
\frac{d \vec{x}}{dt}=A \vec{x},
where \vec{x} is a vector and A is a matrix. If the problem were one-dimensional, then you'd like to say the solution is just x=e^{tA}(x_{0}). What about the multi-dimensional version? We'd still like to say that \vec{x}=e^{tA}( \vec{x}_{0}), but we have to make sense out of the expression e^{tA}. What does it mean to exponentiate a matrix? Well, one approach would be to use the Maclaurin expansion for the exponential:
e^{x}= \sum_{k=0}^{\infty} \frac{x^{k}}{k!},
and simply substitute in the matrix:
e^{tA}= \sum_{k=0}^{ \infty} \frac{t^{k}A^{k}}{k!}.
How to compute the arbitrary power of a matrix? Well, supposing you could write A this way: A=PDP^{-1}, where D is diagonal. Then A^{k}=PD^{k}P^{-1}, and arbitrary powers of diagonal matrices are easy to compute.

Guess what? Diagonalizing a matrix, when possible, is a matter of finding the eigenvalues and eigenvectors. The eigenvalues form the diagonal of D, and the eigenvectors, orthonormalized, form the columns of the matrix P.

The differential equation I began with is a very useful one. You can write many circuit DE's and mass-spring DE's in that form.
 
Last edited:
Sorry for the late reply.
I was brushing up my linear algebra when I stumbled upon https://www.khanacademy.org/math/linear-algebra/alternate_bases/eigen_everything/v/linear-algebra--introduction-to-eigenvalues-and-eigenvectors

It basically said the need to find a function T such that T(v) = λv. Here λ is the eigenvalue and v the eigenvector.
Aren't we doing the opposite? For a given T we find all v?
Why not the opposite? For a given v, find all T such that T(v) = λv. This makes more sense in my opinion.

E.g.: - You have a vector and want to increase its length but preserve its direction. So you will find T for this v.
But what we are doing is opposite. We have T and find all v for it.
 
I am studying the mathematical formalism behind non-commutative geometry approach to quantum gravity. I was reading about Hopf algebras and their Drinfeld twist with a specific example of the Moyal-Weyl twist defined as F=exp(-iλ/2θ^(μν)∂_μ⊗∂_ν) where λ is a constant parametar and θ antisymmetric constant tensor. {∂_μ} is the basis of the tangent vector space over the underlying spacetime Now, from my understanding the enveloping algebra which appears in the definition of the Hopf algebra...

Similar threads

  • · Replies 33 ·
2
Replies
33
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
Replies
3
Views
3K
  • · Replies 9 ·
Replies
9
Views
2K
Replies
1
Views
1K
  • · Replies 17 ·
Replies
17
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 4 ·
Replies
4
Views
6K